var/home/core/zuul-output/0000755000175000017500000000000015156272121014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015156305172015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000361102015156304775020266 0ustar corecoreikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅr ?kٸI_翪|mvſFެxۻf+ovpZjC4%_̿f\ϘקjzuQ6/㴻|]=ry+/vWŊ7 .=*EbqZnx.h{nۯSa ׋D*%(Ϗ_϶ݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;F}Zl8T*v (6pk**+ Le*gUWi [ӊg*XCF*A(-aD~JwFPO7M$n6iXύO^%26lDt#3{f!f6;WR.!$5 J:1*S%V!F([EbD]娍ԹiE03`Cfw&:ɴ@=yN{f}\{+>2^G) u.`l(Sm&F4a0>eBmFR5]!PI6f٘"y/(":[#;`1}+7 s'ϨF&%8'# $9b"r>B)GF%\bi/ Ff/Bp 4YH~BŊ6EZ|^߸3%L[EC 7gg/碓@e=Vn)h\\lwCzDiQJxTsL] ,=M`nͷ~Vܯ5a|X&pNz7l9HGAr Mme)M,O!Xa~YB ɻ!@J$ty#&i 5ܘ=ЂK]IIɻ]rwbXh)g''H_`!GKF5/O]Zڢ>:O񨡺ePӋ&56zGnL!?lJJYq=Wo/"IyQ4\:y|6h6dQX0>HTG5QOuxMe 1׶/5άRIo>a~W;D=;y|AAY'"葋_d$Ə{(he NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'P'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJBR_v'5n]FhNU˿oۂ6C9C7sn,kje*;iΓA7,Q)-,=1A sK|ۜLɽy]ʸEO<-YEqKzϢ \{>dDLF amKGm+`VLJsC>?5rk{-3Ss`y_C}Q v,{*)ߎ% qƦat:D=uNvdߋ{Ny[$ {ɴ6hOI']dC5`t9:GO: FmlN*:g^;T^B0$B%C6Θ%|5u=kkN2{'FEc* A>{avdt)8|mg定TN7,TEXt+`F P |ɧ<Ғ8_iqE b}$B#fethBE;1"l r  B+R6Qp%;R8P󦟶Ub-L::;Ⱦ7,VW.JE:PgXoΰUv:ΰdɆΰ (ΰ0eTUgXun[g, ׽-t!X򴱞_aM:E.Qg1DllЊE҉L ehJx{̗Uɾ?si&2"C]u$.`mjmƒVe9f6NŐsLu6fe wkىKR%f"6=rw^)'Hz }x>1yFX09'A%bDb0!i(`Z;TyֻΗ|ִ0-6dAC5t[OM91c:VJR9&ksvJ;0ɝ$krogB= FYtЩOte=?>T&O{Ll)HClba1PIFĀ":tu^}.&R*!^pHPQuSVO$.KMb.:DK>WtWǭKv4@Va3"a`R@gbu%_J5Ґ 3?lm$K/$s_. WM]̍"W%`lO2-"ew@fϓ{BpY]Q4`Iz_*2coT'ƟlQ.Ff!bpRw@\6"yr+i37Z_j*YLfnYJ~Z~okJX ?A?gU3U;,ד1rlJ#wՆ;I|p"+I4ˬZcն a.1wXhxDI:;.^m9W_c.4/!~x]y7D7@t邗`unn_ư-a9t/.9tTo]r8-X{TMYtt =0AMUk}G9^UA,;Tt,"Dxl DfA\w; &`Ͱ٢x'H/jh7hM=~ ֟y[dI~fHIqC۶1Ik\)3 5Ķ']?SؠC"j_6Ÿ9؎]TTjm\D^x6ANbC ]tVUKe$,\ܺI `Qز@UӬ@B {~6caR!=A>\+܁<lW Gϸ}^w'̅dk  C 7fbU{3Se[s %'!?xL 2ڲ]>i+m^CM&WTj7ȗE!NC6P}H`k(FUM gul)b ;2n6'k}ˍ[`-fYX_pL +1wu(#'3"fxsuҮױdy.0]?ݽb+ uV4}rdM$ѢIA$;~Lvigu+]NC5ÿ nNჶT@~ܥ 7-mU,\rXmQALglNʆ P7k%v>"WCyVtnV K`pC?fE?~fjBwU&'ᚡilRї`m] leu]+?T4v\% ;qF0qV(]pP4W =d#t ru\M{Nj.~27)p|Vn60֭l$4԰vg`i{ 6uwŇctyX{>GXg&[ņzP8_ "J~7+0_t[%XU͍ &dtO:odtRWon%*44Ϯݔ\Zťz;sh4BΈ l8f(q*72"DB&&-ʈ'cC<,֒J2)p|ݛwu0{ѩ2ْM4tޖӳM\Qe%*?vQ~W  yr3-2+=Щp!k2wu_~c9'\ѻ|y7*nD4qL~`|%4Q0q["< HK'f dt(d/ZoQ%o}~Yki7}SWekk̗E\e'hEY9[Nj?7:0@Iuʙ?&Ԕ8e,žLG"1lͧQѶGM]}yxZl 0JM"d.=`Yƚ^"J?}>8ϵq\FOXƀf qbTLhlw?8p@{]oOtsϑ`94t1!F PI;i`ޮMLX7sTGP7^s08p15w q o(uLYQB_dWoc0a#K1P,8]&AJ3DA0ʄ4(zTUWDdE3̻l^-Xw3Fɀ{B-~.h+U8 i1b8wؖ#~zQ`/L 9#Pu/<4A L<KL U(Ee'sCcq !Ȥ4΍ +aM(VldX ][T !Ȱ|HN~6y,⒊)$e{)SR#kהyϛ7^i58f4PmB8 Y{qeφvk73:1@ƛ.{f8IGv*1藺yx27M=>+VnG;\<}8H0]+ES,n?UU{ x~ʓOy_>?o>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQXV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKdĹgJ8@o2k'Hr~4Z(I8!H G8HNW%1Tќ^?G(" 뭗R==9!nKErHc1FYbQ F;v?ob-ڈFalG*rEX}HAP'Hҷ$qM9(AHx!AF 26qxCdP!NZgҽ9l*(H Žڒ;̼|%D Ɖ`Pj . ֈ,ixp`ttOKBDޙ''aLA2s0(G2E<I:xsB.ȼ*d42I:<ŋu#~us{dW<2~sQ37.&lOľu74c?MՏړ@ -N*CB=i3,qjGkUտu6k Cb8hs&sM@-=X(i7=@He%ISd$&iA|i MiʏݸT{r[j顒x.Ƞ"m@Hy_I )j|s#RGI!dTKL&4K>#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhhdEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^߿=D' ܇q>8[¿yp/9Om/5|k \6xH.Z'OeCD@cq:Y~<1LٖY9# xe8g IKTQ:+Xg:*}.<M{ZH[^>m0G{ ̷hiOO|9Y"mma[sSbb'Rv&{@6; KE.a\}:<]Oyve3h9}E[kMD,5 %sO{킒 8.K?]i/`׎tp NvԻV4|<{H@#*h{Yp/E%dlh\bU:E%h@&SEK [ Ƣ xg{z%ǻViX~鮦w35QE~qp[ʕ@}ZL! Z0!A⼏q)[f &E1K3i+`JG P/EG 4 9LڑKL|`PОnG#|}qOR{Q|2_tH߫%pD?1%(@nfxOrs25rMլf{sk7݇fjӞh2HkeL'Wʿ}Ƞ%>9cSH|cEyQp 'ˢd:,v-us"Iidw>%zM@9IqrGq:&_p3õB!>9'0LL]M[lwWVR9I5YpVgtuZfG{RoZr3ٮr;wW:͋nqCRu1y=㊻Ij z[|W%q0 CJV٨3,ib{eH7 mҝ(3ɏO/̗-=OR\dIoHZ6n`R֑&#.Mv0vԬ]I˟vrK}F9X|FI#g.Gi)%!iK|o}|ֵ7!ېATJKB2Z/"BfB(gdj۸=}'),-iX'|M2roK\e5Pt:*qSH PgƉU'VKξ ,!3`˞t1Rx}fvvPXdQSg6EDT:dׁz^DjXp͇G|X5Q9K$)U?o': .,wؓaՁ_ 3]Q16ZYafuvrq^ѷQT},!H]6{Jw>%wK{)rH+"B4H7-]r}7v8|׾~Us?yWfv3>xpRҧH-EeJ~4YIozi:nq Vq8swHOzf ̙eX-4`TDGq G.tݻgq74ŠqBFf8 9Fk Afq#ϛa$!qNCJ4bnvB @W,v&- 6wCBjxk9ᤉ ,Asy3YޜZ4ΓVYf'h?kNg?҆8oC!IMo:^G10EY↘H:L@D+dˠUHs[hiҕ|֏G/G`' m5p|:9U8PZ7Yݷ/7cs=v{lLHqyXR iE^1x5/[O6rpP40ޢE_A͝ Z5 om2p)lbp/bj_d{R\' 礅_}=\:Nb{}IStgq$<$ilb)n&  $uT{wD]2cM(%YjDktByxVl巳1~jpd1O9Á%˧Byd}gs9QNʟ. /ӦxbHHAni5(~p>/O0vEWZ nY3 cU $O,iLacoW1/W=-kqb>&IL6i}^^XpCŋ݃k-$pxbڲ&6*9mg>{rtD)wQ`pkKyt1?[ˋZ5NhfӛŮ Qu8Y4?W֫/&W˸~%pqq{% ?K~,#/0'NZ׽Kq^ėSJ6#j8GO[ PCbʍN^XS&}E9OZ]'t$=tnn&nu [}Ab4 +OLuU{0fIb { Oe\0zE|!@E " ;9Ώf3kZc7B&{u:%lXZlvwohbL_#ǂsr_d >04SRm+0^PTi-"] O('@BKD6 {NmʐzRj.aQcb^CZ-uvpr CѐٱlGNzIveca=%1Qi F>wTLHUGӃ\sA֎Xpljlv ^tSȻ \cPwίwX"{>9V0ټ_`#U8VdTtD_GU9V ұ{q:ObUi7s )B ۊZlzIA4S#x,T3ѱ ԶJ=rs>Nb: Q6ˌ߉J%.Dl2ȱ%ܱ&6XƟ6qg(USok+Po$lwvmi8W_VT18V =| ub6QWCnY'"*aN08wuSEAVخ m3 o\` sHc# fqT .,ŀU|⦍߶/*~48âF,#[:y_YIpʼn)dk!J'Z5=r&; (y*b*O_ULT.ÔD[%s1,jЅ@k0Ցu֯dtKl$Y5O*GUڇvI`b0ο0~oI`b#FOf_$0!i rS/wvҍ%Eb/Ec|U9F-)L)ŘF`U:VK jeFrԋ7EDYpԽ.D\dNyj荊EEg]bÔF˩ք%EGƶ*NX)Hc(<|q@Oޯr^3>Uf1w;mCja:-1_k٘%VbZ˙#G6 `q+MPU~l!.?I_Pĝ"] rT [eTr؟˰ ]\ h! v˱>5S1px fnk}sRmA>d2UAkؖvlX܇Bz1U_#Xӫ+al H d\k/I,k,ρ|`zR/$@8VU^rcG"E7\qtS:ڝUyy >Vc11*?xYa8U`Jw/AcL~|;yj8TR#s"Q.ϊ/Yrx+u6*27fǪC%+A~*Zآ'ѭnۡ|< a1s\ T5҃FZh?EV"sd!@БU ^p%pO3|B5=2怕nwRqR9~ i±za+HFNi>. EWz:V^&YEs5Ȭ N *7{!fRБBSۘ† Er/IGU}APQT]|XN X]FbKjKdO U6[3TTX)|*H'2U0:VunBl  `5/@ա06VNO8VGON@KgjyK?Wq1egI+ I.*F~L!Gf"LD&U 6tGd#fR*c ^tSLjnKS9 Ȼ \ >lr&}+̼d"I va,Jm_u)d靕َ| Vw85F3Liƙb<;dM-})C?Fw*IJ_3UG'+¨[9| >80\+ xJpΕ`p~mg˗%F Rg(6=/r+%a>w Ohght uЍaRs ^d6GXAf?V_mW puȇ S:tŴvŀU#-*mZ5k5r)_x*8ͼx@(k:_TX%[paRu~}#Ѥr %A%`;MxB[CzR怕#H% }8@*AM.SEhd,rKrʇ)br\+! s1CtӒNc_:F*`Nv;ogQFa2V%ZniE|nZ&-I,t*ώlo Lhnٓ'Xm R ˍ-~ά}hs\5TT%~am.>!LcoJrKmqvez܅E9t6FZXgsreHhlٷ+ [}r:̓?W~e6>0E8`Jq-(ed;W¨:Ä&]䒿e;0:|$Ȃ1L-%;Ƅ{dɱL;V[bp>!n&աIJX1$9;[?- й vRCxKVV+#lj@_RL;IQ8ŢΌXD@Z< (1ZRÜ:OUM/vư{'jYXE4S/8 7: `/ +G\ U>]B2/n2=8) B gJ3bcKo̹ʇ\B~Is 2sO/I!}xV&\b<9$4Nve^آ]$LGF@LjKٕyzH 31Հm-XıUXF|\A-2) ' RG6h?āUŔyj[j_ӂ~ яA弆^bDyzǖQ8`jXbsK?l58,?YP5䜭ve9YFznTEf3Ja\,@2,?WYؾNr<V` =V[oB5!Z\ļǪЎr8@*ucѡv\[|s L-+y{5K@dzp`r"mũɸHNd"yc Pu>x2;W`_VR<aӗ&D<=h-Rר|/r _ǖھcߖ]G@Ն;UQG1 '3Jە Q88ASUȿ!:WѥLf21;d9OU᧯MR3V:<}xXh//T+coY5Ȧ4/m0NE(G2[+G~H'5ipӘ͏O +Px SPp.,?Uv|$]o۸W|_*[ԃ 8i NgY,JlU=x;dI]7np83"1F&FI XJ'[JP&9Hp6E+CU|J5KCՅG$p Q .+eQJ=vhʱ4UQ*pp(|y )bԬ 7͝,TGSmȡ*LGUg rxOxxLOL 㢚EZxxˌe?B&9q_ @Kʰq.)[aÈD/Q"}|vr۞^W"N,<~獐?eqeD8M^W[$#nQxndsfԋNk]7ɺX1'I9_Ee|n]m/j;CQAțlq|w{w!uC\ gy3 %tdWSx\Aa@8o_3ЍBa⃸Eg;iC3K$gt.PD:/J6/KE_c_6:<.L*h cc'yI ZFӋ߫D`}w"冥E2}`qnv5pVjgPݑDE2_]"`8(vJ\.pS. v$1 5) 1HL23*x(weӌIdBP0-Z *^(,DO@%i`C[e5{9g'K5<,7*9e (PxyL!gQl.  Y&8~CK?NiQs;4OV;8H"Z9SC{*p56#DFvV>_׿`I=?Dw_*׼W)*i] 9]X*;F G} e*}f0oOod| /k[5Ie =4ןY)"2B)8z@ ރQ}> (NBuܪNY8ǪP(ɅڞTnYu3kz;aJ__ߟ{=2 ƹGEm3 $ɯ7bPI:K6fyL2U#ep-XI^0.hD⊑TGɚlfBD*QW"3E8㣘%RY0-񑕚I@HNF Ir͖5ez!~k0%ZQXQX㱱p,Ȅg&EG +u;,y:QﺔJU3I!k84)ՇдUe,W$hhTjr|=@ݭ0 G8RVU|%X,_cD)T;MQ Zpc(ZinC3hjZ<{ E$qX,kOXc::xZJݗ31U^nl_no[R]RvSgӭ`U99W47R%D ?: 8@% Z'6H5#c \c:zcϬhl6~샄ƈA: j!9ocf86N_Wry|!]|.U#o8+ug^m)n] (2,(5o`BB<.f6 e̙\9vV90"]?O2FuQ YCJ65褼PuؔڊڗByHR )uj&G\ĬIc,9G"p|HcS=+a_[K$:_z_udZ *ܔ¨KU{ruߝX 3255D^m8|C, Vc{~3ٴ҆ey)R֌nik:)爩<ƨL&}hAaUXkVJ՘~wPKXBg(:aĥ҈L\PjĕҡK +s&K &@*jSXJ>Cޞ;eo嵻xG<ؒ:͡-Q=]\<I1Dtq֦b~CuȂafj=]"׳ut~z淕lh}q7V<@s] -tfGs9? 4Vm 9TDdK?iL"ZYu6 lk`'VEɐ[3$iw 1;T-mbyEu<,ҐUN9!m}M0!앯KR*qTM_Ź.E}=U /Kǰw_8;Ζ+0*&!zʵٲ<8϶^ڬ{W\+QmcEjsWkæx +"EpVI"!!ߙLƢLV,t{Z*V <-{7RHC-:1ޞ#lATm9"EMS;7m6Hew'CC!Ҽ\~ggLM[L=]"S\tt1]YڭD N5 1X>#J!֒ԹxP߫f*r걷L^ϨWRrHTmd#ŀn=ݦ%NY<5Hh0` @$3j0Q3V>>Ұe[+E3X`?<*36 ſ9\ڶ]'*,q<ۏyv80V(=wж#c7ipڛo  6aiG5լĆTG#'?KDTpAjhٽu:W"3@҅_jlt::dYEGR18d6D˚`ot-nx|iNZԹ%QkKz]GLw46Z4vy:K.'v[@tϠS uCHn<=uup}[w۝v[dyAhٮV{V}tX?m8v@Z3 oC[g+EyZA3 S.d>.$ qGm\|tV98ŭ!!jI8xau6RM{ VU׹: , (TA,C(U,3zR=m-Yrư0g]OqkҜen[Zfѝ (]ǥݎm.߱e+QE*]c5Mkvad7:}:[eĤ*,dTl=#Y[yeԘ"WtO+BDȓ|Fuu!=̵2:%_E:rNKn-Q9|k/"U";fCq_5+}m}dak f7~UJߢ 8UNw9B#Z6 ۴2VJlC7f0>BA,S!{S/E(r HUCD)\р8ҐFG'}hh9XL :E ؘ.U(%{k{ңCW;ķ7@D!{y&"h~5` gbz==P|~wI̧֬sgRoݛ*1bLK{@Πy(CG j }qK12?T;y͖e\%mޠLeѽkYSnTz R T;gjw`YvI,ܓ^G;,sh!:J~ nٴ@&I] oH+3XD  98^IH|l}߫"%;ً6Y^v]Mnn|{RӇ#M>x_#0s|ooߝ*yFiz+# xcw샐7SZ1!2eirp|.?\Lgd66[*>Pfd r`qt6kf|EW=x>V ?!@KLJw}SQ>Lz"-0˫&UB&=>3)&e[Tu;g9=lu6M`>T0u7.!~5%`*{^ 'ki悧Z_wZp[jw!PWj˟ON-k)"Xhʉ"_Wkoʙmo|M/Ty?d_3 ֮U ` >۪Ş#n$xy(A;*uBCHv@iG]?]cә#3]tRP< ӗ@^wdeWS8eҌ#+q8uKW58S3ryǃc%'2h56ㄈA&~mp&l9C@EۉHPH`S (+?18[u1(xiLЁiu@%C6vݳAZibߍ,EUG@hҲ cw[9h"M ] =,^W2 cA`EQJ"ʛ3@o1G1tF(~™AJ^<30j)AHRuA׵r3wŋ{v=XLz*|>%Iֶxڍ]^[C.y)0[,+3܄?\m}+Typ,н[nh[r ~' '>$9˻rEC],ux rrsj߮G1rt$IB:3ũxFN TGs83I۲#wɸMg>:Yv 3HXn ($$Bѯ# Il[~$>eGn#,ai,T qwֲM䀈$b&m{pgpq&!ϑJs<§w> V3',R2!a6s{6ja!l3ѝ"%2lӒcژPT#Y%PsDWpaWfyL_-7If_h )I2pD\&,1ޑYdTb kVSكU,X}K7 X ieԹfy5QZepFe?%+z,SpkT ;rJOq?MLϭ=PhSG^KN8«Etr2̎LrO^/p5vh"ݳPyw+ӴT7oѿp\6IP ~; "ev97Zۊ e̐tqRz lZq@ٺge8?W]_}"kX_5_~Bw13>Y}Jj T4.ZTIg}'#WQ'䚒BH y[vBͰ;ڗ>ܓ'\LxD(8SUURF=ay uWL:?azl;^IbF=&7X'Mㅃ$\'6Guz9g۪@Y{yyH"v/!t<I6{)ο+H+OV+WoydM'cШ٨$CQKghIV[;W sXIobs3b߬"VnװZ##`(kM-_kTހVbegw8ֺ |Ohux2 V{֛Ϩ [YۏFm4:m>uو{=16b6ۥ+ܕF,q=np a޴N +a E:EZo_P(ɽ#Zz&.Ug6QEv#r_pOL$W~R8E0!O =5K5_ QWEV7A8!0,j] ߦ]z ~W^8d#7{E֦k.U!~6XH]m3Da^l|hid@[كc_겧0i8H 9g9w0[9[9]urvdYfD$!N!6AVPަW#˞OQ[T;m X*6_B*t BO$nC!ی)} G>uXVz`٧ cft*l bHeOmA(_%oA(ߞP--*v#T1U>PD<KEK.8E"X$ܳPfs?i?,GZZ SحZԌQ=` Q5GY?W!Blfh5E|Bi<' !=^*{,|rZzeUz5P~9tK5 PvZ[)|)V rD*@|tjImYُ=7V;,&W=] s`ZoQW]a'4X&%-W:>6z;E8T^z;k'̟Zt1+ijl{tѸ~?0+w M"y`[O<,17K&5CaUPIo?eyvMx!)"@5lbk /8Y$دlz((:|!;*wC} wEjv A[e~4`k3*B6=)i$y@3`p~{1^5z0ЍSwp*~}ų~T^Z@yG OT*mXt$vbfkph,LLe5r0wʩH7Æ7s5WQK!yHDFJP.e89M=(cū0nϱUV=l'e#j)ʽOd:qsD>PvIa܉W'&痘/w>D?|xtq|c*uetfڪ~~18q-hGɋ۞aLKg w=?y^r3DZPcDWy9ɛ$Aۃ4f`4ILƅ,x++238fz  Zd(6jXT$Abn;v9+M ii!f xETK ި:ИWh_:rL[݄R 9Rn{ݗߣh7*_-/ERA[|DFJY HϬt6.R7BS@XfXIe]x6w5=̑ 5c<_sJx)U:CiN-0??3߻v[ċ4q`]ĤG_K+U殺RV""]-cD0Q΄'sPy гb3<FiP B^%JՒK959w§VV\-TG!bگ.-i̺sr`6GF6m{$icphZBkLސ,j0bd/OE*- ߷1딇\0dKw׻N}9N߅mRqoԾ7Ӕw9dx7=¾=w>Pmw?acq~}fۯ~?Q~|oF/w{R ȩUn!+1k|F~?3_[a{{}3?8J1 (.)W(ca2%vZE8J"Z5!Z!̌1*Q8$ᨲG%XS2JJ"'5GH&kOđ w7 QIͼBUZTM( hJ\Q5 G3/ Qۀ+)(UjP68΢j`c >IІ=|mѵh4xrSˤPװە?P@{> m+ѲoT·U Z)x5fh1^*8N( % POŚM(bcbgїe܎9U'1ZB";vFUc\D/pAE^y{t{m2AFHj e׫Ŵ@<bLI ND 5lT>L" ^G6acZHkg#_m6]er(ꠒ7 C*~&1k86o}H4laO\66.H'Ҕ` B"Ax ͬhŧ!Ʊ8Ww {cDWb<[1 ܀/ -4kMtv?eS2q7ljMЭOK<<%0b#Y 0)X^;S=_}tZCͬ <%+&Tdk r,{QN~YcXN|]:!5#zDZWr{Ps0ũztмG޶Kx**֯G)~E~~\ )B #֔_䬜 \v[_L9TQq^) Bx:%t$ %rlMʛ `T.xjx!{v1 >]y-9B1S 9&Km#H:'8*?6[+,0YLTPA} oϦcosbczBg8qc ,YhE"#bHgt=q)vZzcf! %߆t S a0Hct${F:M#*}@${j7x"]עФr(\UPyCZFGM1y[/ 酚sRx>g{Ң ͪ`H/~k5 ;``e!4.gTkl66sNS$Ž]p=p|8ܚ-VLD!LY"D$?>wx0ݥg0R4[[W|p%(/p# $7=e$(b:WM9WbVվ){̙uiv]Z!*RGp8P%HR􊙰51cCԻvVC'â!ǷΟMvqpץ c*q9MI)(5#VnV80 o^.7fIMa)U2a<{2 UʗKFX45/IΫG3J!SIho^gt>j Q$uG[+XdY fk Z!c, 1H:~M *fへbICϪNizV=W<=Rs)jevڣ ~]>ļQCofG`rKiAN:( oOK|*[=) Κ=pt\ $ :5;-ά(h(g8:q92r$1Esbg{Ѹt39WA\-0 缜 i&|#p;zV#[{^UEA<.nVӳjU=!P!u,V<1} $J GO#opkp¡ne]88u= mrG?և@P0b>줒WwmTw-ʳἊVsp㋗eKMIg$V_E]}$a=ok{e43%"M{{HB\jN.@ =2@=?w E`~KO(oGD b " =8S8xݐx4 FY_HN ހ`އm%H:q\npr}Xus@L!N7ԭ|@^Bu(38β1Mh)7EL.1zۥ GUI"YN ">'Qج;nFZ4V]h OG*1\z #ΌtM?>? H1Vi׭1 1 kg}.2E9W}`”cc;dt`~!L@ocdJzfmxo`D r&( -L%r\^kS5O\jJf7o7*Sx ߗ,Ae+Gsb)Mx)8:vg*Bm}eur0+å %%ΖK[8""?O&]>%fFglF%(j"B 7eJ ܇~/Ϊ$!ԣzc"OG/]NLyYII襸҃X$1fv:2UYmsU)b-/R|0acKԈMQ8ΚCbi,.VaӪP=xJ 7tt'a>!<%.-%Tn|A!{]* |q3+p`45p!H:_>g5iDqփS. λ4ZW&d/~| CBulVڌ+U0N^CF!m2ub% tzZ߁wmؼs:Y͍D(ѱ~!Q2JT.SCLiY xQ-DKLtwΥ.8^M27lڲ]0E ڢ3dI/.YXcoqͣ6ϊ!{@һ24spqu̴ ¾g6T*K(!YK{;Jfѽ`g,d:ARmP$U(WԞ.$7;GK/scu1f,A7x ٥1) !&Lpsq*ijjl]~Cr vTc8xb0EaG_5e(%^U㵲~j@$ !x1I5.{aA580~wҮCh (UZUvxD/{۸`n|?[n`)i:~)%؎JcI$+-2Y-oô2mJd}O;v hijk9UjJvp|diN~ i[QV-rBc{װVtJ)$~m"#e 'ȗ+iaB`P1ȤFԮhZf'd=r-؇6 mUvJqG=Wl;1Oy6w!i2[-.? +\Ҭ{6LgB>1V,iеUm'~]~١0?':h0y^@08L/ hlbP9Q>#|8asv~<;,R4L܅uRxz@;7g_2u|`x*ճȌ^'ο5!`A6hDQ=QbuQpyՃv+0*vyR(N+p 0DT@Arzh_/k̵l4=dž ˔[a?qtetF7n|ugz㫛n7\݅C`_` -*ҫrxuA0t '/t74ŻI'7l7Ӭ<_Lr׈\#{~v3Uۥ젅?T99l]^ylQj`VK-]㣷Px3Eэ~I\ǣ{|E0J#0 ƨP>ՐW|#5k8ɭx:((#r:4J}Oxk~xI|?9" \Ia^5^J65~ cQZ}Mf623:hq/7'pHYs~ۃ@Bgj/<9FE3~ Ŝ1pI8m] FP2Q|^xH5g7OuV4Nڎ^ް7XDŜ16@SF I# ]V$L46q+|G;\Iek.1PVթɢl[5qӵ~9LmL~J417fy=xæwQ1@7kgb#ȍJ$V4؁Y[; 9y]]|=yHBӬٝ-~3"93CpjýQ> cлc0un??̍(f WLڒ7afB _W1Iǽ,>lR6!N:?H\.(0z hѤ#7Umq8UUbOvVaUd$D;d9Ioyloq8)sP€jn<6vKՅym dz4Frg OVP"pkkO{]/Rnj<>L0 _,p0k>̗+ 2mI-20]I.(r"-! ~FVn]Km4zZU!Y-B2ۣ*$kURI-YsgF6LrRjdy"}1F76]qTL}S*R9{Gx:FrhRMsREβRLTL3Ib=JS&m M{`=hZۣ|/F0UW t&= }T2j1i~ f} .Er@\#))-h0>Sc4QϙEϬч׋rKYʰ:v?u`(G0t{^ᇹΗ'f嬡N: ~N%*,!~1ԟxP˄ POg'Wgͫ;|KPAhT}綬/{߉eg{u슯a1>EISPGRX[) Jpo~߹-N}Z?V?~ކq6K&3-MD0B KM=Q'*O!ڬZk6*W@6 YyDwύ7Xh .l 8Sܢ*DOn{xjrj hP~j[r[94Ļa0F(vI;b*-d͠1A&<ް{Q凍_bjԎy(//tTa{qToWf_Fp.yHJ$/a{o}Q_[ "LگIyg6}}NPhWV?$*}v]NźuSDŽXOPHΔ9pЎvYLY`A#,פ텴Qj`-§!5Cx[d2 aX* מv_ک̍Y{WU\5|$ =4ᄌU ,R~j @'&/?g@G]8DB"ZB5'GG0ՒЃy+Ժ "߀WozedZ~c\ ;>ȉGg`޺05 kg W"p7&_NI # &)l:NiܗHB;w{i){H[ 尿2;͋4nO߂Qgb-m^c};nĚEXb= J6%݊&0_gv,IQj>Z_=EW J-L.wX/]0br'd{ys'+3$'wykR^N.Ν?||0?{Q83(k>Zj n6_3mXTJ!2* !>\IɾDEvkqjDi>F=3} ]_F|sUh h?}]_5_eI&*S4cgMƱX%16XkG)KөQ V((Ҧ>7`r\˶`8E R2DIv˔XX XBw0#5;# >a+}{d0Ih4u&kc,[$ xK_gslh`*"blqF3~&O֔8tS\cG@o;bog=ƈ5鸢mgUMp)[ۄ=M&mY&6/I,8DZ``*KLabk8x oHAmJLelZ1hidJ"4ݱ#+RKZU[`|d U/1 bQ접cm' 'dԉAuҙ 鍯n:O$5|#OTmdE-_dj  IIIy5k9(0N+`ƍt1=|>H:b rizzl1uU %`>r*$Z`.::bpFF+ !HD7O?{Ƒ ! Aգ9Gd ہcz(9W=!`H9-S:n}!B1`*3 +йzM ˼/sd]ӻ< { ~ɫUƍ/rQH>1A)V, e2 )A1uuFbʕܑ[LhL g~<.;\:9c%>*f Z$!KuUB]"2zt.Rc {rVou.BRFmNƮQVdgJeRfkA+)ʱR` kZۼrJZ/( 3ʞmfֲB嵱Тr *+k\CA9/PS>]/),)dԲv,TEs}x핪*2@tNv[K }Ϊ\)O`>޺SKet !U.Z_A%bDQ*Of*n8JDJڠTM WFJ)k0t)*4'ކԑhn"q~8#2En(Zם`@L>T (e, i]Va0te d],Q[TׂMu*͂iLVnn RLXƃktW\i%kpu(#7f]Ki 5;F.f+5P v-|lL4]iP$֩xpSǭ,u .Yě6ݵV$lkHƁȠO:wXypS̻HTX QZOCZqquvAǽ"D;'ŧ6{ Wc4֥BxhIqrq~zq>{BzؕtfWʡؑ'xW{C s}AqQ?h7iUz9d}=n@1*~*}UY4 =;9wet-2+HnhD3.U8!xҟ)=צpQ++M6C`}Y.InmVFo誳*1{H"ˁiKUl0u˃k+BxUAɪ6Eh}$UڹiYoӠ.Z#GGG٧fiMyfz|6 1Q8.gFa"&/4_3]nv8 pgXdIic Rbfʩ?AuX`O+l7DofZ_l(Ch&~C#L`i~"|\ V^ e;NLw86B x?#أF!`Mpy9bPKEOoR35zok_7VAxvz}XQTeZoOpX]DCĊ<(O,Lb)Oxxh2>hCUM/֛?(wKȌ UgWO`E;sխ^sQE,/ pO"Vd_5(*8ߴ 곓tfԦ!V]OM<0H)k]o'|ϟ<??AԟڠѠ>DFy0j2O}JzY0a B&#pGחzWLV5kAš "f'aQ4P]u7jtcP9nu6ۦo^ h e"NukSl5vwSLAs[t&i[Jb ~7`tf8^(4T ;:S,OzKU /do4 ~o1LV=F)/y)}5L-agC??ыbg:>8GoOe1r?/qQ}yaضqHضj|/gw5?-4_a By/LWSP?~vrtzrLN޵wy8&{>:^)b1.E_hiIĮiF.O 0HxgGv3~EFY C2IJyeuz4* c7ֽwѰ>` BCbyXkRژۜTնMZuklr6on?tVo>ݎ4bk;yQ%" 5SӪ؝\&~P*Wh¬=0*rߦr#.J[a{%Qx}0ikLAygÙ6U]K~ڽ+#];$<+;ՕMu .Lo\`::#`%!#\0i1kS4t3[^Qϣ}ݟ'g(ocqg0`=}88 jގd#}ݡyvއ?zHǝH1]K|zW4 Amw'otiN:uhz~C昻ao*CX@)b"3SdK cSdΠ 8Hr`Z+iG!<8 uR͈BtdҠhlnb"Mф}n"EqKKoӇ}(}Xko^iA ;DwD+XLqJ"I(:y裣͍8h}zUw.><>b$E.q4ZZ|*ػE.ҫ8ʹ~4 S󮙚#=5PԨ*QBʁiIR\Hږ5$MHJz'HS͵tosCG }$ ޶mm(NW'uHaF:i3vv%N&J 80aToN xSRX4- 儥ؕtR0숡zr$r.,΅ u.L2FqgGaӀ|)>:غܧM>jO^C4 zR=kiIJ$WQ7 0q_Ė-&&%QϣtFӿS vͬx Nk絃mb")km K}7"9 %>PcVVi%Uȁ'Cad$E3O{cR>_<Y.Cź냫^ź,|6b+KG2=i0>QXL:9_w9pIʼnТycpV/f|Z;_W9}DHא|F8}`#/XĂUo֥FԮYoJتyK[Ck D&W7[nVNqwxO7}9ϗ/qÌZ: 3 Uy̤Hu̙iON?}KdK&.krr)97n#E8}qsg Ig| knkq lw0Z_4cު *ZW}]E=OW\(lۗ7ۛ 1aͅ3o"9.j7YizDϰau-93mbc/EJ)i2H҇2B7&aU:>$`v$a$jZƂiIiW}&Č$Y {\'«C+`:܉+lHCAP Rw}<ǫAeC";n{ ډGޗ5hrnf5絋HgrNU)3-P#VVX>3[UVE.e9dΉҁw2`w\f1_]$EK23(jS%ڬqJ5n!5кYĚjK{xP1fȀEC6k( ttCKZvf) ^ՆH A m_,MyA }hH#Yh6gc=+RϊNV$PZ"׺Գ0Ze"٫=Y^BVLM{.$޽5XF8~h񑳌DxzN(oE\7x3} 5 iľFۣjEG# 0AJ4P=H9AK:B`I|9LәF\#"HX-m>\pIm7#9:1=תGЇVh=Chp]dt$Ss0JB)1nZVzcQF*URpJ "e 0ѓAHTʂI֘fصӛ4 iryhx꯼ [GsZG)e%`庙Gd;BٽuѸG>x4kY9CWY}XʦZKC_%/j_gQ;FV b@7ָ3W߆OQ>Wzex36LK״mo$7rbE2U*8>mCB+Eds9ztg-6jjdzc)&Vab;y.ABUcbq'n*߷!-둶y՘x(jl*^0\EbW\'A{|0sY̹z*rFo1W9Q4*幊U-8\^%#v;ovnL,X K n. 17ݨ& )&aI̷ 3փviirO} q_y=ʉZd 9yww6dכX E0̫J(SGZLȫJ(TB)g sg[Fʹ =3'29\ǧdsF˸o(0ZRtGRtXYtHE):d.r5>0c=؇v?ϖкu/zh$S\fb]#oRȇ;OB`C\Y-VLe.3Ϝ_=fؾ1krhh0_f7`'.4ph]ؐ_z >?F-x'V\T:c|lrA}j^k.,?gCКBGt 9N2'#d/#N}Y% @vv=E%:QaƲӧ@ xHpĖuh-|9c| zTzt oa%toW_¾{?>0#7u!et4x8[Nhg,,hGY̎,eGsL(} F+Z@(T>S(sW/aM<7Ok'%1D$)31cEA[#UW Ńy臆9(Bܙ\azDyV:K3ܘ=K5Z݉|PGDޣpv7<47W?] `~6mMU[4e4X{FtqGg7{v{˾Ϸ݃ԵZ]xXo,=3xs7\`|W6'% ߏyDp` !ԌCx` *Z WπWstل^L@myZB:5%ހ^5[Wd*ԺV'wx{3 yVUGȇ>Ǿ+ƶ<"fv5m52}Lt0AONg ؃;;7\:m;plTDl3FԿ;b.FNħav/)o'SpRDڭNYwQ;cKۯx,ӷ3WNkҖ"&ƣ{Dǐ,}LӇypv)~/z,xF{~ '9A3gؓ;ř>ႅB3onƃ$949l&m0χq(lJvVʖ1c5:z'G4 s8!d=U䝩cmj:woC(Q3GDeeTKޅ2J_>~j1؃G}u_m٦Y\n nFbp-Q,FqZztQAuq 8j!RyMe UU9WUoi^U8pPj.Ռj3VN,ϪwԴ rAAJ ޸(p$-fR^hGT; [ꃪ TYbwfCSSA^8-f^ӸjzEJ*q[ޅ @UňŇ$)-_UUbwfЈjX4/+`SzջFkRS>]b.e|S֣dɘbwf*ŦMKEr+TϩX zUR i=ZM)-F^U7*֤)&o1yk0$e5^UeU]1:^UuJ$e{Pԣ=(GԔxUQ?]۱2%@t{7n 1yc蚗ܢlmk~o{N?}|[q;@1 ƾ_?&|n))s|l: ZNiq _90^w7#v~?Au w~%[=~vqyuMMx\ s^|?equE:ִHVP]i5"w,i׭x[oZ|G|;b|7A7*u](v3Š9$*AZߺpѠ]#//nWڨM]57W-q ۴M֞kߣnB維DZж95*H_ǮqeցPBS#hvUKH[WdJm Km p/$jRzp W}M7vXfQӃ'2P;Ɖ7*a-EPH- &=tM*viۑœ}ߦ:>U(Lnn! L Y U<<6m-;{,ZjxmϽm8-;hi7PP!614v˄VX{qC%&dz]MAش L}Ok -`Xdz'tG aJ-C>[ YjRKydRϙRV9Ŕє͋Ű?= 9h b`0F.0d'Mz8"*Thc tt$@!ۡ;5'bRj@єVE}SXqLxrϊ}"!:^sFQ[R[TxtPDIT !Qdk@8(w\וTXۯHpi!Q}o)aʼnbk| +Vf>QX1*a~&qdRj8DJh*l,43f4Y3 INƴbi~ ohL֑ <5j]%imF`[k- ?ji?r !Xvms R՞h&50\V(7)d5C 垥ֺNJ琣QG_/pZF?p =$!ygծX|yq |T:1eZ %£zd.|KU-|9<_n*c{ZA%ƢKݤCM%v/F "q gQ$Yo o bްXWG˽c񆀝W*8XcIʹ-Ѳ:LO?+_X؂bfw"JT J@yҥ j]SN9F_X`42a9"0p. FBV#F0gzAkx v܂6 5CKWJr)6#V[nAt?_ϿV7RQ>(q !Lxw {2)N] CͣZv½xa2%ŌFEh{Ai%wvx?T Fg;&SM&R}_M7%"zD-6 SBn;H5m4ד yGtR"&X - b^o%$ֳj%8`gٺ^[)pH88R;.j_ 6Kp\KUsLjl4]۷8WkcX׶V R[ꎞq Lg0lI`'jRZ`~VJܤȴ5r5&aѡaL\$Y4%rT=H\4r:<[aeJ@-bTT@07_ƊcߐjONꉃA$A]v9ֵ 9{|XopAE"b. SBV)rs%'K <%J8EP1w=$cRz!vUכ8L9C6@IBVrx&o%AsMnb2r I̘dJ)A])rc!Ai(>% d5@F_,OWAJ\%~>(IdWhUK,ao5)d5"$ [K:/[:㔼a 8J L Y4@ k+(x)4@啮xoT! ZBV-0Qh9KmXNQQۋ"I!XOܸIB _ {oDKhJ V߅w,Y: *9ܢ*>c2HW\NJUlTY#.w~N1*oSho8̵/9ՆJ9_3^7RF\ }LkݹU ]˭#Rl)M UmF#k|OЅ鬰[G|u^ujXg ܃0dfj:?|)_ K~f&ZˠC_+0 $Ӄ1|'VeѮd#k=]̩"L=ѽ\ʐRw5cܦ-Zֱ~gZ T̲ BDʾ^G[8M/vX0c0è1X^ 'Z}su]7 F'.f&pK;AmX칺] ׹-|S,PoqQ2?Qݛ)^2)/Wt]Mj_^kmrӕ<ͦ;jw+09a2 XLK<}b ~S=^yvwO*9gukBt͗Dm6N4!V[ʉG !(N,.vGd0BaǾ9A]FSqA 5%xqY>X8lVJ!2sbL^lŶ^lžd/HZ6R-2b$^lŶ^l. J9{SLU{kVTv\@ +fO=*JAO @B-@1 Qps9+gǹv| +-:Cwk14E1a$J M&e5R03L42)RLp)n\F3B]&$b\՝, qJmzJHw6w '@=x!#֢qQo|IVK/)"G( kCQi8J08Pg=Z9./أ#.AMYL'4Mi>~3(kܦHi"mS/9EJB'> _>M,)RvQ)Ro؞^"mSm[MX)DrV9N,;)DЊ^lŶ^lžd/-cU23"5{ XU]-ǃm\)ʹ0v: ͇ׄFKƳt||v- Lx4;kS,O;_OQ#T%ǪI"oFq]0qNS}ݙͽ0h2^/_cFW*V#Y$Տ!4̵2 sb"xI|8]LAl}{29&tf?;;d*t¿K B/?,qkutߍރZc% VAOV*sP0D$QBO.FUBId$$p #ZH)/cӑ 3al!K#j=1бdxGiQm+4L9N6sx&5_YMͅ-oܾ6Uʯf#N<7ָ%od$R'7fĆq3j#44ttntF~5fئuatNFYj&`}eaImvIc5iBRkaczOyCMR}ZЏߛCJ[e{s9fyWҕۺɴ:A-}wP^x,XR/9HHt$ *мh< A}57ӟe|2߬9u(7(L*bu/F ·*Wj]'p$p"j9րcԁ3lH&'3lL0g + =CDRK<0b&[5TJmOJ?}/ךÁ'_in:ɉ`!O[԰Ռ\O [4Z E-lH11Y˱Tf!© Qyb`lD.&JEő_w$"):& A`;!&> snf,\C\5:J9e4& +`h*)>ckc6-Fz sy0ϰfb;O<¯Wj} 69.7_}'GKFjWTg fvfͷ$MOd9nF'ȕ-SF}Pk 85'LVSIQ懖1'ං77uU1sIAVqـG+;@hdfD->9"m@JT$qSouɖ./%-?a!Vydn\J 2rFs|dVKx1͠8rxVOE]}ܟ#I'+DKq֣b*sL6zF\arEQH]-Vn|VOOƅ_+J;JrDBӢ_4<>O쾂0:nF0Jv&yo؞'uicTLSnD&ϺSnFD'N4$zʍyMHlǬ}1BaA?t$I8=f. xk~ؠ/<,Wf0e- xe cC)[J8_0n49l xzPT90ea0Ytco|I_-?0G} Q? -]2"[*`e2wյO,Kޕ_W}>'jmRffQ[ϩ;"ܟnjnvM@p<}o?\Gܥ\V(qssS 'oBFZy JܛWVAgW}MD/{`aBލ`p3VP ZE0E(q.~wÎ7+䭔*b%8xlmʺxݦz=W)ܢq%hJMb>(g*<_j 0X+ P2N,_2ЋP|Nn`=|pf;5i)L&n! n6H3\YCk;, IH5T`+x.bC|o.+p}V$Q^I_2ܛ50!Ƒ$ǿB~#UU%`ak?q,â>OHAR[_dIZz1ۂNGWgeFtWTiivI催ڨJ/ g6Wkr 5ʳ`1Ӵh],T Vէ Ɛ5SuO|E+ܔ3HTg*P #%Ġ:*D{jP]ju;"4F%z./V.22`gՀEFNkapr/up;Gmqk8*AĹU X.*C)"\CI 5!1J8QE0. -c64ˤcwճk.kʰP\L.+QYjXkY B6PB ׎]S`P|8+J'SX|/b ~Iㄼn+)jWHՌլ@&4: $$X$+4PIw`? TTAzƲb2 VS ,+hW$1w\֕QAޠB+~XF2n`Ɛ)uڀaKV6$ @ mG"-FQC(d֜G9<`&bDb`.Lƾ(J A&pu>A% y_ `bD8+4ho@b8U %E5n,T B##yGD)$h~GSmHPSވ ebUM]Q^uì[]Ϩ`[QQi y( RD"$Z*KE@I ,+aMhޣƻ },|p.#h< D+9E{?qrǬsDq"p>Q )pl^vNZC!'D̿hY;Lp|G=+jzt2AA0p1-7T4gҗ-*w%SكҕjJUAİ CM`EBBgąs yjN(@dUF5 l5T8=Ѳy@4 O×Me :[*2(28pGҗBE,TG7A`ZDbYxm0tY+ QHk"S->wu4S M[ 4"1`P.O9sA/9І(Tʗ蠻%|̡pB0=SANB~,hD-hg*I)a:k^Bk۝BBrFPi-;Z{f ̓ D xd!;ybE,$ 22ߢrVEQm~+FT W ڪ6<:s;y/7]Oߐ/VD$K@@ԭwP7ۭ4'EEU8'=໓*٢mðbYk Q^y(Y=yp4vMfE @y%o毆^fĞL:=h7LJxKT] [Sz'ȍڪR[hONJTtJփ`XgRdSbFHzD:SC}[] V"4צSkNnY!ǘ/z7( dZ8pE#$qnlD;L ;_CU],W!BmT62cT-P$o7& fnV< ZĦULm<3i^'6 4(R6#TK!sOHה#H/{*MFn&yfH\cUw&*>@PHyJZE'Xʙ9T!-l8EOzȕs!5=?H(Aj|dNԋjOQ I9 JR0W ҭ0 Ȩè3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:רdٚ:AVǨN!ިQwhUZ0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0<32Vk2ৗ? ~-F5Z?w4nu~F è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:ƨwL'W{?^JP9i{{vovY<&I_cP/fZMf _mΏN-2 c:F+ f;v;5X.VZ h;xG(̠8V;մ8J+ [ X+JBiހ%ٰkwkO?ӕS+kd] X"҆VVnl8R%`sn%`=YKk@ٵĬÒo#X Z XCq%`̚\vkYcV }l`6f-1jswÊtsʛ?mY`g,ׇݳpֹl.O7wM[m|_߽޸ Tc'eU Nv.,M/xQEd42ڵ$o FA_nϲxvCM&_. 0>뷧'4LoWVl7z)oZy{.͛l.3~3MGdU)[qw|GV<~_;g W1UU]|4Ӌӗn?_Q;kJzq bQXFK#pd ] Xm9` g\KLJ޸E?H^keoiEkɳi޼;'!ӆՎH5 *W+NC߃mO?ߺn;t<]2Dn'&}h`{E1Ḗp鰖4_K;ZF;0j7FGkv%`mW<^K̲Q%fXX:,7>X=}Ǩэ>QA>lklaXQ.ب~!R}XfGDv> ?VYM8W`9LsY+7&(Syd;ٲNK>]?-߆is3YuKyuޮ_$˴fɩU`QV=;c>9SmJFH9vB aoVG|hw&z6q%`)z镀eMqBoLҾ$^|8za~Yxap7]X'Z2-߄A^c.wSfVk}>}jnⅩ.^1Oݶ5[#9[s.&A2ݔ7w$5Y/>k?^^~n]wiBC3/zNggnoO\R;|78tv| w(z-q1~EBDev-=DdlM@5Yim]ҠBBҮHG,]f6xT dGLO(uw5r ={#Tɒ>IYb=uV`ٽL}49e0F5PRJUSSՔJFvaoUsTOҤ^ ɕZٷ*BI*u#X[1cv <%ZC%x4[li%HdQ/^eTc'g &u}--ƒ»zyx(RWNqu$~k!0epwQZjaa ba$Z,TT.JImHԓlAgyor DM_b|f^C6fnmQL۝3km)D ;2D{(%x +6mBX,XˈohlCieR':Y4`,|x`aZC*e?pZ(Ϥ4  *DɕY$\%OU8ܪLzD)h@z~c6PXgDg19ULQXPIĴwY-Ұ1k s12Qm8EB$j$gsEΡd0(lx*` 5OMA[/U i)R:K& U Rbb*2%<;6 eBA{H;hcu9 J W4"Xa. Wj,MvL˥trXF/`B]}ΌɮLߓ8U{CJC+q3F0ͷQ$SH@ LhmJQ?jI#5Fax鬒9&|0(l1IeSPI!0hsmT66JAdՌZnj?lLWn_@bd[ePQ3F4GRFk,YQ=%h~EBYo* B꾺2ZՖQ=ppR}^R6"1bq3EB¿*`H <C?FjA2a+"Ǜ1mVuDMOdh[ cW5ukk^K7fG GcD%e$gQ^!ô0S_iSr_ PYS_w &MuD /d@X8pi?0K6J'. [6bxLF ANŗXqa7k $Pd"e+f"r1 `(*1}OA `((AJv+97Vdnh >:d,:;(*`ɉw%,pa[PM겒!v2"}ƒFon[kvM >wy2u1qWC EW4 w`܆Ŭf\ZЋAZ]%_GW1*#Dꊺ&(=`,Ca"r ю2ƈ;fCd`:Ո *N6G]Aڜ׃rF4F,ZnkT5>pPK B;viwR t gыhӖְFЃA2ءvGyc""+YƢnP\1"xሳXeH!(r",Ul?rXNoH,YfQ5*)8Y Dy(m*gU2+QECMҝ orf&= ]$/q7RP1aH56iol{}@I(K pquU  Ipd*صLLvMKQ#bԺ! 58EMyHY5Xp4x;fvg,Y,o>(5w80)Q^"o*(9Q%rcl90pgdt  J`Q* Gdi Sԃ0V6Q1 M@ ?zV$bʅs urp嗀4R/yMae è ~ʢHbQu#>j &ޟցi,e?{DmX)+M;'#j pNiu)fndj?S_ 5jAX/=byPdҧ D2)I12{Wy~wPJ"CzI7@KvG 8Ze :rfF8Àe%3J z΀|Q =`e?nDrz?͠dF)0 8%-C#V̯/x@.* ks !$*Vտ.uLC,cd0+&) T$"~B%N p\N&!KklB9 Uk"!(Tv70Z%XP5aKh!i]ōr_fb!N6]4[Řz^9E]2l;cAEz#믿qy>]Q,žݭ/ɄVt(lvzUt2]<ݎ.~Л.ƳяWz\^.]\pa2]̫ٻdL@KDPڢ`fs2ޔrvyE6A?o | -^.Ƙbwe. ]VfPoT kiFE*w.R-tYT뛐j}2$ l}PIpkl3xoB3+2Fk%{BxW \y`gWdgW0su.pE*ኬԮտ \#^=eܘGdC4> Ww3f}WwRWp\}sgW{&[jF=v"+kp •B ~FpE;6pE(-Z;\6J Ô:#+2A WdJ\}pE ^lhX W'>k8˫ho__gʕ7l6e&]T(LfÓCXz<rm:Ӥ=T_N*.WA.c"} F/i~ϟ//Ɨ.Lʜj2ovWH~bTIٜ|ԭ&VtT&}:5:a̝\j9z-4> 1ģՊ̧Z~lZhn';كiv&aOt5svmaMhv|ym!gQpR/]xUTb: ZDŝ8@{y}y.ͧLGxCLǻv6l5O1b)rV]rN]l<7jNJ;|ȵ]CI(eO Sϴy0=~;r>,qTRxS?@Θ}>)]C?,=j~iN0 +wD$ݟֻu38K"k: ̔hk,ppvJGPFKq{ ob_S\0L,NX_]/|5YRGQ5w8чjpx6i{֫}}hoY+^zu@})E}wFB-zϠ6-}u\>,9)F[o*1<_~UzY6?Y܊qM?qG~r*=ėѬ`$$;dHp'-4j?|D,/~lq1MHir:h5'wh`W,v1))@45ns߭cz7:ɩN{s'qw/3J`RکZ}4sC KQ^d7n>7wnͧoX] aCϚDLsw.\xK:J]u5GH&k\c~="jǦi$ێ9:}@]gju7=wݛkoC,Xb'tŲtGQ#cXɵswSQb |xu1[ S+Ц&^xcNqq\͹Ud[sHN$1y(uF÷ 1?u {txy|:TP3F,]^wI()v/sk~.?S]n)Xԗ0w)pN' kK`9wNuPW<)w@jM1U%@iABiJ\Q QjoxzW.1i&}RupB삈1&g&۲刋-G~;8}zGL;kRn7aN"йQGCq#",!] vٰ!gߡI6gHę=դDQ#q([DpTjVd8Uiu"[GGk)`x“&I +9+S{&cn1>z+sc)~&Θ g p"cQ~:h:Zv W5y?XZ CrKfZ8+zΒf#ibP2=ڈ'LxQyA19*Ƀ&ϩNSqCcj+))/r'I^00jH*"9UJfCw;G輈:/ |(Sw<)R 1dZFq/ӝ{sQx_n{'XxyEJI"aP^'TQ+Im\ӧȧ06ƼAcʻk_k;:hL$>^,~ U䮍T kfmJ2D1X% PL$IwqC٬C4언93*ך"!B1W\Iƺ,gqC9kh-LͧTl]DE;%r8~\,輚:&a%O F;5G >jDi8$t™\4׀b#.n<]).eCx^6S-mJ \(F9d1OG thrC2b_fJY}+gOŒXW齹\X& J)FR뼀}ă]$)( ޥ\Xyy҄j)F k4I F7#RRKU"СLEFS.rqxC|^-Ba+OOyܳ7Pϴt>&5eL_FB})'nV1,E]W#Ղ4 ,qP|Y9-Hr )Ф!P8.N%s}4j8x5CPͥ`̤os!mz(Bs[1B:M1˷-!i,+-|[pzJbF/Up*~6߬=hM>?fN"!ϯgλvGywqt |y9Õ7s[ Aտk.:Yքƴ6Д5ux0Ԛ%ZRgG܊˩u9 g7vKj': Ft(N/u7vc};1b5bDczW9ԒVR(xT%LB-:ݸk}S7Wg5;wS ~͡ڕottMsx_()`B³`0[/vt~ -ܶqm͇ \*3O~[l}eo. <5@yy;_ntx?Κ_]-ϰϫ_|ʛqn5 oOXE{9lHվ +*[zTM>Js:si|AssjL:\+]QXc2 'x>`NEY?,j8KW$k(&Ll1_ηdTV/pg E>[YgUeg~9+.(n;ˁ;i`r?/͋e%ms5=nZv?Dfٶ^ptW_6'cf?:X+Ktr7Z@=*W6HCGw%xpw}(] |+ vV&)ɓu$Z~/~?_̷H؍Kޝ!V.wm/ȑVJ[/󛽮ʣ> u} F,[2g; sL^= ~{f8`D,{g?% -wHk zm@>+}XKŚ_@Em| SZϸ ڸw5^c\u=k WٺZ e,1H&x_s l(PNܜKJwz; 0V0"QncyW#?,H D`}b`G9 5/BR%YzF#6p؞ :rb$Sї"ESuJr'E F(.)J{)'6c#/v4:ANzQjW K|DkD;,J 0V0#Qm#Y{ʍvʳ|'P F(.v#!hYx)6LsiѳST^EHLF,qryaWCaaj_=k}*$KwJAr.H?'_[ZrbBN[tj@F;B|$P&14'$k󯆎y~בwz6YjE*8`,!EDL>e㧏7z"R 0V0!Q&x[/sw`QP9q!L!˹ rAr9Cn1#Av?/n\H*Yi% Pxc`i%&inIЦ;<ޡxpc;#ΏƂ0,bw_vA_4,S"A++5tԻø"E'BB|o4 _cIiԢ%D}]T"f"w$.="! O -Qp5DxPC| C&%J)#~1]B;ݣ Cʰmu}2bh`k &A%'rп\ԨSr+UzwHJcUzToO%]]Xe>_4 2~Q%7[ b ;=7ն\| o뇌hekQ0\BUQԸ$f[]?fPcogP :P˙eQ=?|zv~+ps~_=?w[jE+@^XV<#Zγ7S*[n|9f~o{BK/]U,{>%:C{٪ֺFC  ӷ0ՠy؎|aԼP9`Ta3 BhA$ Fa"#7ېx/``'^Fq&k#x.^; ,6X-V7^(}E{^`߳t8lf1eT߮ps>H\2Z;1E NK `U'r4k1)öGRfމC% q'0"yzw^U~"^UOpeMP,79jN ^,i'~[zzz`q%E"cv3\8"-H,T\pHtF?%ZC7%(^*AJU$jTL$IЎ1D3buN <M榪M&V*|)5khڶrv =M;8Mi."^8%"K(L;{eZZަ9i΅s*#^ kBJeLL9g9G8a,74s}*zrKxf]H{ %0&IAcr qCYqr}豱( 貄k@e%I,Μzh6ƂrOܗnp'U1dFmgI5++PCe "zA]9e"7`9?k)5ڪ6?|aT]TSix}Z.ˣTi2+:hYb9| ˥kZضܹvn}yIn6UBVXi[$M#Q9RK%@Vm1H~/\*\KFBƹ\ܟL;{cpXrLQR]v Η;K52)vSL'Oaop4*dȢPy:@նQH`D/TC |(ˍIes.$HE &(+layqbc{ÀcNX2鐲groSȄK:tJ~T0Fd1:I{UѾֲƎ \Ɉf{Y&H X  4ˤ\f'3T@1#/CUlw1ؠ4nT0受5.AtFqX O& ܒ4[FXo6#:{ZsOfWPܵtM揸ZREQ7Kmi:ƻc72>YՅ,=;ldc#/vj9^bAEL bti>ͩL'^us77Qޠ TNZI%|RJg+W ^8'HS9bыi5>C;g`t{=Z镼.`uJ@cQ7\%0J |4I"RɄ-H' 2m u3#_2a!.wИ2i ,lTdqe=bBpyF8Z@ `jfPcrbQjоEޕi&D[O۽V_G!{&tF \ OB'I#1▩ݥڀ_LpgɳYwۡj,2a⟶k穡G_>Ox+jtZ(",YI8MiiniUf^@(Ab+MV?h˞McP82<r*Y&&eQd DT'2NRvz5o#.UO&Zb{ )9%tH>UubJl#hJlcmRVʙו-zd\d3P#FwtO:fKۼf>40\̈) a՚"m.r?4 BoHZxzlmVTq~_ufˊ-வl-@ʘh\94J-py' |a @;` 4n"+ƿw^}i~X0>hv4d)ٽ څ B 0fV_k𝦋A𴍙FaE O08Y[gf /=\qʾLhtY#Ĥv"qRpQK"UƜ TqI*L 握2p`?9ٜh\1வLjS1^5دGR&t*r㼜!(wm1!N@76bד821NTtMir\V"k7V@o6V@e5X9FO^ifyVF4% &ߤ6lON }W>iw^K X-w?9޻\k"\u_éә #YېI5SkK)e_1|Ł+NyqM( >ːZh=Y(/T>c=x [dRd$f)خ&|Z`rWVڿmNS˷ o^x:hL&D`D-c$Q]m:N pW9Ze0pVzBP|cGoG,vPoiAc"p4B0G-؛ x*tz"F7@Pxֳy}WپfTmS{(bx9h>x_TϦm,u5IoDo rIˇ'ؙDO~-~Z MDDS*\|}##r'3 xfBg s‰_**"^]Fh(,o//rR*qVPf"&pnii%Z8) N[Ubކ\5ӟK=1^)_*W68e \\EaǓ |FܮaG= 1.Besİs"e oP18,ǁeyh(}{:^=jgl6SNI<_%~>#S~5S(ݹ3M(C2SV57CEC=v8) U3:'G^}X<>c|\"&,0X';3B猙!3_P=U=ɘt$u!R \<:9m p4=!(,YB@)uIxEg@2b45I@&wQCm]zTv܅ȧ|Ŭqqs[;$y ށ(G;>c6}(UguV`k-.񬕅ZZJ<pTĐLfL,,S^&kbX.*I˺2u {LG(V38m|0+*# O}Z5_y:@ jPT|&d_͊o$Uh늬K>^C.m 97LkV$I *y Cw7[V6+X]WDsX* j$԰u9g\|Ў5UV(ZXծy5 qB iFtXxA-$_HGZ) se%{gڼCV_ק2\YA%uHr ڒ,8=qRGY3BΏ;Kg[06ւ{- f@< , 0aO]M[ 7dH֋Ug/}懋f~,_Bv皇\vƓaG<%ϵ>~竍$XR葺܍6m11]oڸXP-E$?uZǣIȿ媇RH< DŽ Wc 3!ѮT$}E:ÚO(BjMVXՉW^@T sM/^f0:ٽ=*v,b4sO(u.W&玏e8ѳt|YƩ%w]`Rx^><<|ø`n,}܊Ҋ;hDfΜ`1_E~A&x,bdޝ{r\3cA#2s:s^U{=@9jyG_ ao0l&c:5S)T/CT̯jd%c18G *%OeM AІ ʲL"riK 5HpI>MF4窛]r}} X9)iFPG _P'IQZLmZc--z(-Ib2Svu/HA#2s\4%)egJ#LP؍W\QnT6p}U?L)\3NIEfqZJ_@cPmx)H& (gLp8lvK#2sR1݃3Pï^bgex!۪W]c}E$2{y[CAy<.;D2c?`XiR?'zKu8 *( ܻ+m )WPK.z*{Z *;rP=|!< n$'&.cYZ$1=!`P\]`ʔ=bU"4WF)-U4a-̹QrJg/ReFZ2k.CkltA»$ ۑ4+@o+ P/3"|;[Բnd>ͫyсUFd\erN7,;6RjYj{-gxWW;hDfof^mtЈ̜+,eYGusY@7lJ=xSjLF(C6L!)g%zuEֲZǐ{Uͽ52t勨S#~}En-GQ6T#L 4'ynMrY1T:+7Ov. `6*E q@, V\d)}gM6)gY&l14  w4o{gL呦j/j}LBXaC+(9ᩤΓqGGO 'E9g}`O~^y*.8Pշi!6J_7TNim9{ F`^ {[Pi+;_yt@~m/-~\z+~L{~v|M=m\}ʲXT{,/M ~1v~wlonZq|y߂;/a8i/Ar}Uu`̥sisPK B)SH9KrQ e ƧJxߧ̟aP?fs>_g rU-fsY.n)s@Ww+y twal?[o|7)3™|]vƛ[O܀up|:~]Y9ELkmrr@v˟WًKf0#^y՝֟5dNUfwzlt>dmyW[0wﶊ@nvݶG#-܏e~OXpteORYU-ן?}LQR4YD*_޽pSX=LS:oιJki"_?!9֣&FdC\*NkqF_0;l8MI|60Kva n3Hvm3Ӌ|+J֣m]vKTc.3vEVbZA|EJkT*̹6=x9ac`p~I| ue2>P5Jd0rBuT܈% ٗp>w Ĭ4# ``YEN".r'(K<X8= ˆ+g7=mz`QĪXS3^?\zӠޛskƂ$(pepa`ٰ80nدG= 4|!3~Yy'z  G"1 `qPoKept9K@L&B!ѳC~V[XSTGќXxzI Z(Υxch8FQhwOOl Eu(D 1']ӂ8RzѮiI] c!뚖55,L+eP0CJX'RaђzQ Z(bFn_4YԆ1=7}s?9ĘU|pwE A]O֏`2jiTs+hȭR= sdGg abNUb6u)n""^SJ5j<)t,`LϽ=ˍT%fރlm<|5 8H!Oד tJٶ;<ޭaJ]OZ?/O{m7ܻ$jve>M^4Q'h~{h~eA9m'U;mS)9% 3,x~q1jtxcR%>,W&h }^jה F=/v 7F!tzso;|r#;*8hCGČr=;Ϩԟ_SI`e|HPGr7\):IM:γ s/$64w1K1ѳZx0,;z%.DQؼ&!"93vapRO:%m~4R;DDp\)ɴKT 8E..r R9evpՍ_sc!n1:2r@(5h>]|?IC/-}]S~D~ q= ? $] ۶M(9ލQ f,A@do2;UPTFnc1]jkX_yJpF&Y=Lu,reIB8vqjx}LPį0y_ 5xw.8~79ֿqQLsb1:/xT\t"x4Q q:pJ z֟V-GsGZ e!tWf}f2'Sb7'nQ%y3z?~ L'nNeEIQce>Em92J_k54'?ў@'H )ByPWjaۏ;p3wN=3/wX ,O'{hDuL!`g 9F=~P''VԈjm( =/@BQ2u;W(!~tHn$i140_6y6.124F<6BD qGyNuXDߙ;Di$ߢ#%u!Dk"bl(QS䉎%?R[*J0Z`ѤK/7k塈_ UR$ pIJ,sd "-x|X?Z,%,A 1v c v9ׁҠivE{ +x̜#I8SSbX6cSyB{9c,_0w#< ⚫枬fR@$Оzb2KV˥/H&=9je4|Th>o]-:cfʬWݛU7SHuzda>_OKu{1+38Ub" v &uG\eTkgi̘$uVfc%5ajyK>:}u/LoL/cezBw4= !lkE݀ l4QS pgՋBS3*IXuN"kq,)07DWT$B`yama_"j]L(OE%6SЀ྘Ǚg?cg? =EpF3UgQ фoG.(Ӳ(y2Nwz%}Ƕ/8c*njH2v ʹ`,KM=d$RkIfDKH|阙CR}U^w|<$(&$bb˅z_~swd怸'[O )^@5]$U:Gi %Aɵ0lܭ| [|KkY"O#z0 Eg[0xae*&,%K]OIҘD@X`"^JDJY{fu푍c?jS"T.OLHf-*fX_V[%5*bui)L̓7 k3uF-'LU*B{~uQ : ~3wc3D(%62 4' 1#3 +r'hZ%umP2bFnQFCaS7cu$E"n|;o |[\߭Ol;踹q&şo{Z~ݬ,?`<-_lWe,|Ul~L)۵D7Ћ oy .R[tϮ`7\o5|wi٧0ჿs:yv牛^YOB|?k|5Ifc~<t3wGůד7;cP)vR\j. _.ak[xxU *]fr |D+y T'"TygHM|.ԊQl.K `m a__$ *~d2vJQQ X| ^}WC,>ލ14PƓχG{&TSK3r D`̸OIREi*=@;iI*z6bo"pM$F#V Ef#9ϻMô؀`_mZՉ-Y^i/% Q,>;J_ Cn%4`m),?XJl"'[In֫i)+qzCieߊ$-ˢOdiNTqe3fz wEP=ZDlgors#{Z䁽m\\(E)e6S yb8 j $KY+8r*[̷ݩ*w͢ (k>sf0 *Jd|aeS|G47e(Ǭ̗g4x -YO2_YO\X:ܴgU_lmˮV(m´Dk3+guql].VËx?f7|+/^²ǥGP nO—lnF<r)d/Y8ݮֳck7uQ޼m똯#j{{׆0<}ϽxDCV*x#wNk_O0H#) X 3/'ruʺW7u@ZTM\O~ n6뫫(o(]kp?kv6kP{d{ƑB_nM~wq&l엻3)+IdA҈HMrf(9:69ճ/G=g|X/9=)'晹Z(,JWLK ALZ0au2im^Z2ǝg|KǴD ՉbU 8ý6Cnt\0]㒇KM2PqӒ-nZRa/L?8>`C wZ⊦6鸭*行z&zL 9<"ыHH24b%?xtê`V%75cBeg&η`і\U9X_E8 @/׎mEbz}z/fm"َ55zQU: Gj!z4=Jzjz 2 Y 0Cf A!ޚ{@lw8;SuCW ]CWᡫsjKz&N<썟?dFv~}{㿵rː2 IgojyrnU> m= }bL|l oŃOm]Q,F))J.g#$@-ĵ*82W#Yr0w1?u|?'0)9Y:, 441Hb(g֊i[#cEJ-QSO1inc%ה>9Fּs_L'[9*$a,Z+ÐɱKHQ(45T!@V <8uP9s2Z3Kاt|!Zn 4 rSպĪ n5gvvkj}69}oy4ʙgF:'rrW:ąUw]~(<~QV:ѓyBL-XHnIű{ ;06`mK02&gƀ.7x O'O9 0KJ;m2 kKgá|ho6="\l+0]8l;]<:Ǜo>1E9F?ۭF0ZjgN?^#ߜJ)ݝ,~麟O-XO Wxre[2ݮv'[cpI#[ysT.+so9+rrd'_ |P +EDGf6Pz>k~v/|K0S {~\~TJOi|u~1YOިkj6/eL0GKjy3(}q%^EwW#ѧv bs6Z?cﭗJ3D VňZg.`F.RցfW&qe$;նvz;YI@%1$jV S>Pޠd[[(S25/*!΋ ̝߹U=Y`90ت7~ܔ797MMysSsSV}MKn}-e lxXHIa'J(K;ϽS)_ !hyltNȒ_ !=Rc_ 6 :/ :VnAU֜籠A2_sr\A_JU*5' bd ͙vY}}1UO*0upˊ ʃ9檥}Y_NTcnOmJ-/)e 1AN8,&TԨ>+f Wwu.qGtKfdzU>a{etW{Y(ڼ'f7RMM f"VP.~ٲS<.VߍYohrO^='+{Y v-Aes/,+C[έ@^T1K+&QMP2VK8Y%9~fu+yyW&QK.䳩L]R?TyesAv4]KF̉4g$Wq7#\Z=eqXôC0x1:RJ{ZJXn~Kߙ֭Vkl[}u84p}wUu:~vfn&jzguu_״ix<#k]k t-.@%e6o  & 53qp ϧO`|S.Vog\~e(_WwT#."v56wªF3F=ѢײɖL$ܯI$ۗ/2 i+ sg;0!c*tQ'h8ezKqZ[Şos{8]nZ!_z7…л-I[MКi*;BU_o34{䮘LqWzh+~ TV+r*l޸+T}qW{ OmǛ]=Z sFx=G8''{Ly 7 RRx&R'J|cgb)fp5n8="hq\"0O͆H% JYA`[η xM>߸t$b.Xヷ)q]B T2౞eܶ[d&luN>6 xp]]%t='h.xA spW4>I,-v]$T ñ-;H^jQ|Ҍg+ג64;Z| 5+!k4$m!$ՂET$h)&XoEB z}z%&QP"6g1$Todm9 L $^S BITgc[} 5#Xm*3L8 Hg1<*' cܛ1EB sS|BoEBYx抃 s18[wP$<Α)$(HK"B$ש |८^(הA(dUU֖*bmcجqyox)Ϭf$HF[wP6U+ 5% t%5UH[egK3y0[p 5kW>BTb9ذ`RocK㵉[$3 I9<*6$(Wsh il*Hw?| UO6DԌ g[%3HdT>.*kSBR<䭒' TAS3Je jg6{O-PŁ7 h/Db]$MB[HUqV0ZոYZGwPPs LD t>yn! Ѵ]$/Uy,A CBaE˒ʤ"e˔.wULx2XωwԦHT^ 5kV)G )^9D`H !ZKS.X|W#%XH60A#ɺ]$oχ !Sg F*I%)YҶdu]$o%S{ZRW|PۧTkx'^{T5ϓKc##H K\d]$NUsY$:-d2iIB҅0K։FQHީ9Y6Y3P+qM#ˀ;P.v/)uv)"tsHD'mAvsb.G3lHm QeƊ㹥EB \iII^;))84$딡-"aw0{DNAal9EurJ)joON_98)cv>չu-P[˟-tzP_}7xaݫo<mȳߏ~J>N5,Mg()2 8gfxJУ6zٝ?[<|M|pRs]|V2*fW9i)XSڦdvE3j&ݬ("1cYQC2&[G\.d;5)M:P|<9Q6hq(aV0`CR⼵j-*x6eI1+&>Ͳ0J BO4yeq6E+`Zeguy򳼸N&\!zgdy<45.}_/%Ļ7|J}7z f~j¨ԸwGW;eoG OK7}yy[/~'9] Afl]aZ?˻Eޮ&l[Zs5Aި~'d||tkHZyzCt% ;n9'4?9()kz"tLI\hb ї;1+k/ wx/2kh:/s>ᕂ?'eM $.-kwk~z1`n1B繶sjúi5.g'+9Gyq F E>WyVyNT߼]O c\6-7.u 71^L?ֿE}X9Ʌ޷ ]O?x O9ǿ<  y2;n>zgPL۸-<\s ގf絀8*'D?bgpSV+uP%qF)Kǃ:Χ7c4>oO/f0:EA `W/=ƒ( .&I;Ԝrksk|;%~C(OI#etp,8^4H'$\)%?#Y.bĕ9~nD?xɖHY KjP$p|DR9H,ZJ9d9km,<88#*&`#F~I3o`i~m9NOVbt1nw'ypYxbSy0GI fpO~:HXóW=<֮͏_FO_#E-LmlMi};gk퓇1NnBX^ygv#[A)-vq 6&-}Cƶma úTwd]=%{=/%d(Yf{a-cʐ.SY!] %Mm)XNNfR>S)F%SO]2q ӇD5iaU S2NOg V*STZR6Uv| $=/vŐVVWOz יm7=g9pR4A}5M +|h3Dh;rZ{vNü$A DSUA~ʑ)R%K<#Bd)\%&۞=+RiX{χ@*3$1+he\)Ø.)YDQpQROPM{,1o…! AM\E3gb[Jx%b3QJKWH2kE4m1_oj mf199rt/[4F r ?{=8/ c,BWi1;C[wM~b} vI3֬^ه4مvvxdRQy6xx!HX0Bn\m =&1St6e`.M9vonm*mg3<.xR\yNO [U|ιK_i`}u{Qw=X&~vz L&t~H7(*j۱]Տ }/e[Vٲ-[e%Q\KYy[+t'x|F k2ɹdz3@} C7"6~sid2%&W%{16whlyͲ/o}-ɥB($z>)k7y;& #}n.4v2uޔˍi%uК۶]ď!VV[ioWguZX a4;yjr'ԡ&5\z%9wC[ZPg;&EܻWX($^dK$y2wu5X+W(bZKbmj.ը:R-MOkR:wTڱТjMگi_w/WNŪEunˠ ؚ+9X-:hOml*{B a 4Lq74#/d餈sکbľ {V!=oݿ7(wȖ%qOn?ӆ+r7u j:>)Fc1RɘӉw3ɓRC`/%Яu3a0p:0ҷCo8-h1oq?Ѷt2→&nRE躚mP7Z35Բ,-pAtk|"4#KcgxT{DQrE4p8E`å,hoy޺D0+) D E8kTx2hϟf{wϊmU_\-ѫ>H羹KoXsw'~!bg埽uS揸~oN&1PSH_|E;APq=Q/ q}HߔEݸXeI}%X[ (֪b?ɠ|Y#o:ue/2*VDqq0a\w梸.(n Z7,dqQ7yMK#w/nh+ pr KtvկTrnXo6?>AUެo !ۿ7[]J⢻ӭG΋O@O㻡s{~Z7oOrD*KJİx)dd=Cu+wW=R~z_q޶RGHkWm܉쓈lSSZ*h0{҂rgx;VH9C i' i3J D8Q$ "28p&ơ,b6uN4j-yH,YFMS&]R)Nk1N I0g.dacIl@'BDPDmL (y~8_y[k2h8b9 (.9WJ_V,o+wޞȨ̚<3]K67 6H> o*mߜ&QS|)kY5jLꢚ&Qk??:.#b+j; uk7m,)pWߵ`H79Su[Uݒu!WFOw͆\ӕvTJsҢ g Wzm7}1w/$䨛 7+XCW;]yݒIW"+⃾P_A;`: _~>b&wiH4pbm]j r %(߼~w'"^^<W~^b8/1rogʄPq$%5(X;I&9WԾhr:X>.@BQyˤL~Krϋ /GY"g*Qύ&LyW -fBq^*£"'zEbDkHM:kT0 DlUr`!GJ%KN{Hl-;lnEiQQR 2|7Q:(wیι6|V1  r>>t܄ɍ`UIn?lҘ6)^י5<']F~ 5GU3K0 +,4>/$w0M4H$ZoEldI@[Aukع`JbW I{^v*N_X-9C2 kuޔUESŇ3P4s]궼颸W*Z|!īNúy6tmƢ}m OAJk+ Y)jT&nϞ븼I IvAK#Oǁh2{EAz\?ѷOROyDKIGB9T\ٿH;߄:{BR|TB `R B=ZuP,BK^`*(YHtBڢ+Pʣ=&ʣ䨣 +YÇ{yjƦ&E$gMO }煉LsA/I^a =1͇h^i)5Ck 9J1$B?l1pI0ss>ߎzFiQ'D3ADfgcOPL1:'T`tFmb5EEC%˨iʤPSSDcK$%\2N I0g.dacZsb8KBZ'@u.I ")kh M|g]rNPxs2gM\d<~뽥M+oK=xM "GS,08A%8%A8JZ2 :W5୔Y5VL=;3kuۗ/ْȠphkH>t*mߜ&a>X|3c0 '{Ʊrqm QM*aV(Ec=I(DX0E5MyF;Gj0 h$,c)j-4Ʉt}]F%U'6`dfYˆr XKQy"3uưX&  f$I$G!F.YO!5i.5wJ ]qcrF\!`v>.•1WM7WRRՙ34W NeU|BJs*ts(\t"s驛zZc2+Dɠ3Wgi7v? y9)7ƿ/k:44 }VѳKo<5v&}.`0 bM9orבZ|ۛ:L=B!JQ %No^}![Jȩwmo1$i%v}BLwjILNd4h\>}_c4y<ǹq<'ApW/ E=1yhV۳n he:C%"s!\IL[UFIUF ݚ9++i{J1W-o|tQBWtJA>EJ Ba{[] DW=uu QrE;suJ fD*hŴ\!\- Ѣ<4\!J);2x ڴKӞT; b1MEWrsiHaԹ rZsuSP+zzS  ֘+˩hhA6\e+v|J2"s W&hMTDt )eDZ\!\C-*h(eG\1IWޏ'JAU)q٨䢷ϺA`lA`DҀOS|oo,C# >xEIId ƘJp&I)@bGɌQg+6|(..f̯/1.t}׻]zw]YuM]?$TlQ$nQ=']-ix$eςd˂eRY71\!_->O c1,4hy8)-:nx 7??=p0z>ǗVrnitP1 FfN L>h)( 18IJ9qCRG!Ľ(΃$eTA҇*&hyPL(IaMb% z* MD+whG{a=[BWJ*}fJohj21$BK\~RK_uow*M\TPH&FT !J'ᯔyNykQ9| aio/7%Xg JŬEk٘ m9M:F-|+&KϬC8-`8k`5A Ȃi @OR8XAtc5ǰ\F=7Qj(@DHOHÒTѣߗ&&rk4c)~)1eh!OuFݵBM"3ɎX#ͷ3M=_pe vωBN>|(-0ƳT$-#2jToz˵XK\ߧ+A$NosW"0G<\ 2Bsq!<*%R 9SKY*1a֘ล)IUL,s3C#e ƃje!2aeeh 9 @v].}+Zq ѳ" ;R["+[n}EiY_Ywy} E$ +Fb&Ɠ03=kU3Nh19 IWd_(,Gw "]I\n°#yP)G)XH y3ZQP#EEjJ6ܡX  8U!CYH8RVvjeKL+= \Ȃi_D)QFz$FIOL''ѴsMQc7uT"mɸ' WfI呋;ַC> Jn1!Ȝoz/egLqvI󬣳|Q6? WW87_?b1̹+ /Q.?姶O-Rr.+ʟ7ϡ"mqN]92˅ЧepF 'A!٦lkvWK-yJ8j=!1DTpW?Uglwqljn,J37*(jquDgu}\[~Yj{FrER/Kxރ2ǜIe~#C|y}򵝠V J'cd"U(o<mKAmlYlj8`8.sme']gd x Ө u^yeBr^Tj)1;| ;vrYR}h"x'o?a} >KUkY'|f~l={d#UNɢ ָ $,Rt^A*="Nv9zrȳxirnz(M&x ٓՠ/]O{Q.ZzW윾cIXuE(eEw:Ll]i%@pR ү_&qL"X " -ZD XW4m;V8fi@rnm*%+d'=$';NW2rj<+?sjg3:Ӂd4|N' DёkPBPSQ8omЛ-g"6IR G*UW YLe$F(«ayKH c1hЍ R+0bp'bdrK>H A&'u<@bS,T9h$!֝ .&jy-_ o V,XhMrZq !y*FĈN$MFV;ABuX]Nsb7Il k3kG\Bq/\J]IӨz~z $YG'9R _#FЄKYBt*FѿXkֳg[ӄQT|)U劵qS>PoX*mcIk^ҠC%e4y:; -(t*P |~)?7{?Dʈȵ0Hb3Vse.(<8ΰh /`gj/f')*hj)i'_bTe0VkS6 yvclZmut%EyI1%疁DBY$t&'*6AJՓqQ[bU/UB <S|5TA KH!)pHL09uW:7-ЯF J=>XD; D锨m4BhSӯĽV~<ϾB'Z c1xK&NY𛷣o_QȮzku~6(`P?"7Xcd^}my 9Z", 5Qq宄W{_wqtpc:?vMJ{p/NMgXbhU/xU<53WᅓWE{NގozބJ V,~_nʳz[/.we}Xc8cŀr(2🋗Jg%>f(`{aIy]yEqvA&9p.Gr0 合ܗ'zowO>p0K"{߄#tfg7_QɂhwFH%H ; ؇|^%KҤ>DR4k gꪭ$O&S)f$3cyΕn)sv4 ZR*ùKg? >ױ4:39eE>4w ѐbkNM' Vg۠kaX<tyAO:>FsXN+VL!H-Ҥ l&%4DJmDGE:ygoU2ɣ{ L8:ϐNhtr?j:Mtr?*WtmtХ^ XomOzDj;Vb*vᔝ[ 쟑s0DC_k *\$`Pz6Ս1 #ư`B̶_nd=MOhrKFyyQݶݺN֙4x||gSx{hhiB]]фKR& gk\ CM$kwĔ8*ے)>.}!"1]x& ~FcjY%:*bK@TWޱCde/vyсaG+U& `!$Z% ȼ/"aSBõ"f  R0Ş"">!’]^qB`bB`RBZ{P$kC0Jo uŭ/F]KQW箮J4ЪW$g;䵋e]?\UԧIKMG;~:ɮƃq!_81zW0 [yT^A0N\~(y/MOdM9GI}[%=T62ȅz-N)PGrZ/-LJf+S`e1.ȣ!ba(xy~P`ReIkTa4^g5vFi VV6\e4K3;jo7hӼ9=ՍщoEhR]ٴz]k<?#[zo+>;GBpm1:A3e#﬍^8șYJO%y_˙AH%)CpbIc76Bx!dR1qi5p壶e76/?19+-%{k/^ ߻2|t|rGލWv*25F)w_rنǗK]ڞ7,#NfdEbH<f6,pmWK-Gpl$/:}P]渂fIuh6 *qM#Ԇk :kfcQk4&H!)!hT|qcPĸRZ zV~ja݉E0hBģI)ȸ`ڪβj뤘ܶ/_Z\iA"40pH2qr#Yl1[ιߔ.8Ԭ XTo]C 2fFx;&5 LFY<ЇրA%O3iI_$yqB 0wH n1}h)'vqRR_귝KTp7Jo:ոb0v7wJa_Y}R ;:LB?SK/→#5+UV`!}u0*& 5')q7Oox80| !poյ}۩$euzv[>Qs?Uۋ-]M(|iHww5‡rSQ?!i:ᘘ7W(DqMu Z쉰%+q5 n:=rCft.rM IGnI_Ux`Zh+QnʾNw\&!^?3*n@~ 5|݌ oW (-%,@0~5dW49 •Gz5%K*l56Q,9$cɻĬRM@ƯZ۔_+& Ս-Ƚ7Q= p-͂htI*4=´uW*fi&\7Mc7zyLQpQ8sk Yb2s#^=&9I#D- >-hrɊJ2:# w>sɞ۹a?ѝx%ˏ+W]>鳋2!U5/F Gg}T ș4x%n ݶ1fbt{7<} 81Xk؎^ ea/MGfWYdc"]ؤ#]%7Ii? C~ݸuNʐO@_7rP8qCլnЍ~<(O=#ν\_1n&_ŋ~ l|λZrYdLg7n~9"]Y8TlQfz6TR O2 1WS!2Yd\J42'4GZ_M;aGw܌qwx}n[_K8E;۟?zdfv 6N4.y)KACB<)Nd 4x128'~ ڨ" ٨hM(yI'*AČ؄]k4{MfiorCw\'q9z\'.>X:LqEFvG/wjf+ts3Bm#zvm5-::: њFG8ҕEFG #]fidɵ4!ɒ͡,<" q5@wz)B_WiP(kS.ߧ~+3YySw:߾O }w5 i'ksC#8q="v}v%ܸilӅٝťQuk`Jpk_ ɺmA[ v p=4-:B4\i^"Ӽ\/4/Q˟ry Ry_i^-4%h1\!k?>7,Q̬ SDw5oFO57V1LoW%ANXr. !Lf|VF1 bD2NE-بR\I]fyb<;xQ5'HZ4hyz.Te#R;#4Wso1lK۞2 0& $gYpȲL>$cH)dq ZmDx 階M^lI ms|-إh|ۆcNq^  AS*EcTbeVb920ESQ樼hKe=ঽh)DB2Id}22f9<^" nMzN3E#gɾt3Q!m{G/ZצrCXŽc{TN(ϕ>sf0O>x Ql5ْg?b+In5M?' W[ܰa9sV kvI*/U>I(/~S7$B0#RF\$gmd'Ep lL솵qVR>h3XDiNД!KZ-Tj3NX9CoES9e 7=ZXf+fsXP჆oVr GMANt,E23ƟƓwGodK֢&cƪ0~=hCްt&孯}AtKURMoy %j QROf mɉM,ϖ4xKuGeeG])2'4AsKo%ɴ/Cƺ:3 J̸PKkR2D'R("c"$1h()5ZؖiC{e6l4 S]_&? dsj33T;X)7Y {M\QKrw=A+,VVK1d9:"h|ݣ՟T wa7RRq`C'UĨ{AIO+ &@(`XY `uȝJ: gLOq,'ROOwmY4Z~$ 3q>+C )9v{ni[mRG"nǹu.}DͳE~ֽ_lq#cqmăEܔo^Q ̏ʯD*vaxrQ0er1ڌ(+^UۄqEUŢ[> !U ~YFa:R2|e{Vk&sA/wws$qo}9L衐цޤ &+уjMfdq+2if}|Gwv?SxD@9Ю/'kV_%Z={ɢpMr'YM ={:ɋժ.[Fڨ`ڛ[}GS624-j^„2EB;>\UZuViեra*AZ)6{ǕhHqS*U̒ε`24t QR*S)mvt(Z6\G%= bėzԬ4/›m,l&&MjBp- wOPE0cRN›1-.6:{`tFbht|ANI8_pf|oѴ:2dY=ZǬJ; ۽Wh4IUxPIɜ; F o2k xB. :ރ%v3a:Q:GBF5VH#6hn_?^AB ;v݄&˽e;VEσtDhQ)PF9%E bZ0z`R;JQ[TcD&7cGQ3!0xգc&]BqOҕȪTycsjo7i]YqFw -zIXAlRԞK=e2 d2D ᪥hs蟐uN^[i:?T赫-&fH֫aU|dmi#`X©v$Kґi =CL iDF9rG ljrSj>܀NĒ`⊙4 Vo  .: zul)\KC$0r(:ff5$Z!t$9DԅKO]`ZPk|'u ]wrt %= T GV/â/w{qry*w Cgn2gM憎z"K7NIk {N?%zs0J &hb%cT#a%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V*B4 >p;%-Gʽ(0N9V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+ Y;}@J 6q(J B{pw% @Q *V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+HCRpTR^y&U+ 8@NF)ܥ|pQj3 \. [ 6Vϖ[†?-Yl^l;T l4 }X_q'w=+|:'g͌u;Ͽ4ϿN,y;{qz>ʖ(+ YmjCVڐՆ6d! YmjCVڐՆ6d! YmjCVڐՆ6d! YmjCVڐՆ6d! YmjCVڐՆ6d! YmjCVڐՆGmzwxɋm5QI?*$ucP #WGxH4őQ1EQH{Bɜ vG봨yv0/ <_ԧpM(A~?Z.vl}T3QП[Y.O'GkxoAt =tHǛM"I`l['/8(a׻AŞ Z*f4sykkXVg;Bwrˈ|_tj6p{FZ}z)@څGNm 紐):F>v0Q_-_-3G8?!)wC&8Zr+pxݺA+qk g}~>|y+ b7]mU/3NCgo޽D[]N RHzZVwVVޣ/80Owr3w޸\1"ַNuIW,Y.4q1G4جr*U/gSp tBV>ke?)Z#{}_#!t *sz>N˿wBr޽:-\J(-jWFxWkhYm.koOtށY-zR~e84/C6ًj\ݞ /7lxG\x U?}}svm6>p7h߭^ ?Wt __A=:'{zw\}#:T_}!*zԧ"7Gf{.+G.]oHW|B"w]n7&ik#KQ9~$%KhQcTUuuUYN$ph5Hzf:Vr3@5ˏIY+eJ]K#(.paKE9km: 嬐_@ F]ĨpyDSښty4a'$*Eq OPCL\q:hS W>0z#MS”QTjeNT倦=Ej&PZA-7t~R'̈́~R~xqly^δ"0 JGd`zohHԙu,x"YG+#-FAJoL?4{b̲Bj5߬̚Y觺rGwPۦ:m#&f4%XskH ( ڈ7MFfzpD ISEh47],ň~o:?kE^+|eÕ5pd?T#f;jiTS ~ilӉG{1cEvHGkZ -K=AdX\p@pKVܒU̐ݍc|ȯM삚Kgu˃? / #0 Msk=dJ`h.zӏ!▱ٵ*[L1T7 Ō93˭"I2̔Wj; 슆 V4mCO`j[佇w?,nWei:V8k[^@}YSUZ%IUp4gh՞2# bd&Z+IŗuHo-@!L*04Z7F_hyzE#=f7a؛}>MIa ]6u/iiu8z4VZ[v3Kӟ,p^=\E2ëUMZNzA}0y[ \*tc!<{"L3]o5zKϡUjےu/\s AA=M;zo}_&Yg䧧,d/9] gZr9AI;WRQ4^P]=:T[3re^/ TJs7xUirG_o;@ .aV7Iܓgf^- M ^ܤQf ,If% ,ӝWW#}{5ro^ &!F1&^kI28BF[!8  _Higpj˕ Z#sIsdnp'Ppv:R" #:&7U?T7$$"  Lw%brcPUk~ّl2 %/ЈSDIƙshī `\(FR g&͔F+-VD\[~cݰ8tiڗij"HFs"9)1T=QYԄ_&7 I'H ҫ* I26 H+4Jh&&> snf,b"J议4D8oXkH^caR=ˆܭ/r)O0O𶗇A0G~XVĞ?MFe7&хS7*"5K`b)X Z/az<5bXG"`)"wZ0yn"%Uצ+aPHrZX.*LWȧGV}KRE@ayF̾lx\$f$35*(yzIRITPu.-SLp{/ciP/qx:%z]* s?ItW9KWnHE:B:n$x3E2U07Wßu!I5іO"*L?|g\Td#݁]*Rf<8c&`G#ڃ:X`<3L0 Uʹ<䠘Ň?7yLYnHOۊN`0G4u ~X^CԌXF?O1Gmx+]˃)rwNelF0Vg{1"tf"%SKɞDoz94ia&x1nLK/NL`2X>,SZJ-: ru(&ddQ |KJ\MY6G&K`v0@/t[VPN˧n7K1W9L@.lp 9+tf8`U޼|Pqkf)µUZݺ-O(y,o%S/z0n{]>~W_UW~(jzW>w&a^*+x@U-ҭ- GK/ܰ yn=dz.aSo_z{9ss/ԯ$5i:7M2J͘\:Dcsqخwo\pz%#;9lB)tޙ?i *Uu,f5ėM=p"0)6R| Kz*.zo.#}IOïðfFo/nUzx_qͷ kHI5ᜡ]1a˒,y?逢 |4 j-$* %^F@HHLBDN6tCʞE.qVq1 A0 ('ֆ # Ҙ6$YNHƴ (QHLd$&A:łKsg H:9ֆ-;I$]JIKbyz*"֟ݜc]Q?-{ٰN>kja.D; >ʌm7ﵔ:c2"ʘ&3{ˌXmmϺovvko;0?pW^!iCͥ>Ϣ%U}6橚WJbGMG yQyRɴ4E"&CJ'0\0PҠmPV# $0^b@`ؙ5;rCe m&ؗ~Dzgnv:{3}[ʚ̀d7jj0R谰Dhi#AJX ZZg9:)KX*juf c`vt*gFKFP=R0ٻ6Uux~?,`[$1w9`OYgԒlop=3C=|y XÞzʑy[Z@%vN 4 # b)I"D@Ј RATT!H M3Y4繁QSF]vN0OiROHTZ+3GC#yD <] ׊>2RSOvpR{-sZL=~IЎ@9 `jeZnOZjn}-Qh@#P~w9acK" SgApdSYI~=GM4hJNjk#5T$j#X`aVMT(Ls23FD`npYmPB6h$:\(\qj.)iYHg@$TDTyAiƝVc  B' LxUKƽ gqCR|JzFO"J2Μ#JW̠ `\(FR g&F+-VDܚNձ;?:pi:ij"HF\8rOFPMk(FfQBdT,p MnA,N ҫ2%@H^q`%4 œ`a1IGVg!|YWJrH{ާ )u7]x~,wz%uB 2R V+JS>eB"C@P5\ sWw2 ņb2^sWFI -##}yD3]W-"{S4Šnä7\6 1bu}J7%3~ kgB,3~׳oF@ՍWʦ8F#|J BSz/P"NVQ"~ӋSOvx)‰F"rn%8ۘ+3{z5% }]ԟMU@=E-6`MH tHfB>Oպsj~T<,VQM)bIhLi;Jfny$)ITP_)&\=XW1{m4cH^zo@*Җ(8*S™"MAo8cօ&<:)y8ZUOyL&1Hvx%A!ltRc =XQPdFq 묏dvnkdL? HFIQH` DaO`0^^Hl," \v1<1:'.k"ghS K8N=6g/7'~8(JT󟾀}UrBЫdJ.zkw7ӆfHI+b#X3a0HiާC> EoiVW޲R?Ԛg\RRS<ö5J0x[t }ὧ#B&RR`ZH<O $= 327i<2~zjXlD -\OQ_{h:w?Hxz?Jwإ&ݨ4g&%;RfU[b9 2(e`3jX$"a=6MVHK®-6R&;4Nl& qs髛u+"[|98*$5jxT"7 qf%*Ű[)r9x;k6Z]ӄui~k VvR2@놬?2RڈK8Z9؊%Uf5d>Z7幏S\-ի ,RE.8tPf3JV9^=*FEIN@FA'ZK(Iz yI"R:"*Xr;=Y ȅ4*T0u4H&$ZB`#6HcrH؋K'c]i@ QzHLdu4(&9b!ckuIZvւPXWtٺ K>2lx7,vڔ)"]VQ$ξ-4eS+ƶ`NK+ "EtFOq^}xحe/_hsW;q 8O+Lԡ%gqT s@@l E;UJQm(;@yr.ONr! y1F=1ԭM햡*ryơӨ q: $D~ [otvfg7Ssen""˒4yPvhY;cL<(7kA< XݽN6֜̓?{ʫ{{W':/ݞ$csy:; ,s)n o'1NၱXhT VTK,H;};P%wK! :!P1HJ +,YJ]:8v; AaN1F: lp6HF7g)Oc.*bKdvE!*~y.6 U):pW qkmܸ)|TgqёQ*ZCG3v}*@I܆Ikr #@RGTPJm|,DcCF #때4kP#*!P@[wbZN.Lex7Vc=&SAo%,OzoK-3ڸ<4~/Ksv۸qFQ[3 ^UmvHHHJIkHkaQW^-Ff39 *X0B`D$4P C)-1a " 9U4f!C+=P~W?2~lhS |][ZOico.(oӌ^+ =hLvּ:h\v\H@̝rAZG-hI4%nT$YN) ʛO5Zy%JHΉLkGbMQ_s ui++$n>Dvqa"0X2c'~S& JD,#Nr 0|I*ÂO9ѱш/$\Є A2 ǨU1kQRk^jF5rlf봒j-+d&o~_mRG&=گ~x{,ˑpΦ'!&PĻ%H;}9"#a1re10l^AEjP֋YHw@$1,sfEwZI%PD0@ĮD1gcPa*[W5EĺS3爧U\.#3AjQhp Dw/ں!$F\j8rOQG͢&|-oBp b;B:68oM 2FʸڀmGZ yPB30qVsC0s`,zR5|]27lK/3 WJrJo3wK> ɊЏyѻ>VQM;~Mtade]iYWKIJdftKMGKbrpjq5M}cWp4?hX\-b?f$mM3:Pj~TqH4lPa J`awgu5sO1{7p̓YFʡp~׳odɓT6})">JL_$''O _C8Ywx6R$_( 4WF"r5BWmLݕ)poWnyl():mq;!† 9_hܧ_?%VTbgx8ʑU&$6*GEF)DN4T:1+*,p[1\),H6]"U@RCX@~ ȝ)td?|qm;u}ZԆ0]gc8Ev<1d2A*8(Nj,A[#Q\aBudm+7S%/sT[Mm (I0$XoH4Y* -TÍUDEn;r36"LF(5Ē^jv7;Ge o|oVKr~Wwbt+N>s`m [_tV_tA.!1dzp81IÔё=q U<}ynPD] ?s0\{$M`}J]>.XRO] s8)R$ .zH!J鵵 S[}K͏xq;h2jy[{n:R*Btg4)Y/{0 GRoJ (r~D.9-r͙A6ƒEi(].!ղd[ղdղdkղB<cL Y&QBH%UIeqn@7 46+:쉗[t^BTyGMܸ1w;Rk|hOT}% Ƨ$(/(1-x!ܓr$=j5VY&=%WGu5peMal@M9g\EnKQjT,yAH, Z-P9p çwA$LŴi9%%+O"|o`_uI|韬0VΤՆ'{9vyÐ$^a <* s~]`PZʩR ˽)o.}oq\$)l( X* %ZI>RLSNq飶ěU`Am*l־ˡ=kCo(U_=,k9, W]Հç~tr\#M#ux_X:h 10BB2Pg($^2!YobA% 9g H:k6-;J(cRESb]?(Bp_ɗ|not@$ɉ֍{.r') /f T`r͙J\!kKu))}}~햏tnƿB|q_@g$#ɋB۲%EșUؗF'THTRP|pѪ4)4OIyJ SRh @% qj_ud J,䥋^ooYΙD1u}?us".H=z'$ahiϪe.&9os*at4:FosPt6fG~02xu~3/܎Ony+6kne[jPs}k᭟cHN ?: 480=~irlYPQIl@AX̵Q*g er+gIqTeUe]eVe )2-LSXGg)u1q8w.®*;[cR;ĬW逰 ߜ <K6\e>_>9*;[䳩4rb m3NSh'*r줣IGX %&:004ғ$uD9h29AZaLv!;뵈4[ IY|C Z*}\WzYOKdVې.m]ugӍ7BlEừ7σσ`Clף͋Q~|~-*퐶.%eIZ'x%rqoa+D(\!r -%& }'lufӚUX*|jcj|ByYБ:봖Q3,Rg2D˞)RR%n9 Q 5]칱6COa~cwɸv6B!$Ff5r6'+ԸN"ῄHT|> ȟ(O r<Y+냉KMhӘhA #(H8 Q p_ƽ/\ntwl4 b,G8՘ո-qh5Zz0j`F`|9BIY{I\]Rqع'eVj|?Jŗ[bx$%ئ.{>Էs$^MG7gI7GpwJ }ucweަ6 jU ,2PfGa0W4B7 #bnh./_LPz۫k']+igw/p]f+m<.t= (;#.b8ۺ8^:w[iVqה`P.>;Zԇhw:Esf:}zQ|˹{'ФÞǾy޳k"-J uZDI'}UZ,M^L]ox)0P.L*:!Iѯ._7M9xRE3/Ve۫ˮֵ^Ւ=I-YB}$)*L0K>f;'gzH_ tYy^ vv~Eek,jJуYee1)IC/gx}Wui BT9vL8Vwj_kEKO6Z,X.jnmР8PḏRAL!+s̠!jaxe/I5h"GjFׄ%,34ܜSL a#v+02E͛^ u >PoYjd wW4F'`H*F(i_/u?}b2"NQ 2p& m o@ s$:ˢ"/P FU j1O~{gMޟ0G] `!hI&pDE h8L8:"ma9nu]>n&t"~͊9 8}?K Ux?BSF &^҂R¨MfcX<`RLDDo9a%sgDY/ȡG ]_DMQ:II+(.`)IX|M I2ؒc~;(!Z+l{}ˁӁ𒪾duNc##5csS[-D%ܐ0;8+,_UBGLRax,eMϮF٪XԌܻPZZ=>JB;Dh)Q)a U[k4O\oCCrclӇP{y/rMA|{./Bqۋɵ{7zui22i^?jӸH^OG$E{3೾~~/|b 48Wk =P4PO9 =ڟHu .I^.+"C@u:n %6 4q`;k">tT7#Q݆,W _S5hW#:H$k.S؃dOX(5uL'7{s!X)Φ&g Z $reacA >Cc bpc!@47TY?;hzu*ȃf=1KY=2n2!B2g/4;zJ~{bzIJgc! @x{1=`Oo&߃7Z =h޼F=˲x2;,wO^A EٞE ~CE=JZlB[Ly2¶e CmĒCz+`۟wF!+C[)a\x- 3~\oe6mnaV=uc?wbj$fCn߁4]ִ2`$LL)Nm~\ƴYt\v2s!C*Scowխ~e Xmb ey=K JK- VD -j&E2{mQ3܌mQöBSNJOG^/Ɍ$6pu9Imw0 nuU$Wc~Tp; wk6wqWZI]e*w4 Zo2o?~?-5JrOap]O sWM~zzἒ򬀤*>KAP/sjV,_~{_+$%uNX A ]JO,iGR&^r\Pk@Db4H*X`o>%nDRJS8>Q*`h&w8a4Sۨra4SF0e\]֓'8b+ODTRW o5F #*CaNQdH yxSfʥh (h 84P΄O籇8J333 9A:ODŽ8) > `G1N7Շ?78#J !J0e+HU:E$r´QDry j(PD(S> 0aAu`=M@HA2[10*p8O1oqFgǰV_z+=eo}7>z0ƽd!r=#DcSENLI!WSBJm{r (+OkQ'=բ*t4'R!* m*YcAIDJP@S#BQLWǰ9a .[tv=&7IuqvԀYY)jc%(*19NN%cyL>Q]GgC*N'cbl 8 "^#SmK|bA~$(rf1Y*x4pw$S٫d*†) -KEqj1IL*)5QIf2IcT7x{l.mlͪiszpW eBvy@ѳm䫙_Φ_Th[뤑-'s<(%i!`N ÀDU sHr$bk鏖:~^!rnO_~ /?|wƻ񡂇kw]-tS-Tc—r{Taa:҃\̋ESS3&#^ ([}}6ާ V7X4پq՚oqϓ_;nt>r%BfkW8^w~b'f{J%e]fnesO/j7#Ӈ ߐ3yowLlO/\C8'_? .8x0ɑcj=5u2؎nw]KɅb-wxk[NA/ga]{e݌vsvCH\Vf Zs&є8SZQ)b%<+a79i x6+dD`yjD&C;E˪ͻ_j}֪vy6vY2Xha072B[Ć5 hOty=0tU"PgҰ'p[%|\EOhy)meIୃ w")E?ſIZoL/Mw0f^ƽ_᣹->o3{k{ӥ/IU`в^^3XD.FCx v{?^r{/fۍǝ'9 .:ǻE%^'m}%T`+9 `L%*FF)oJuo,8nK!|vۆ 1ga@#I bs6 H*4D;}J6IRz@h_GŽH7RO  Z0䝵Ax'?Bdk'&֎ 4^v tMuCԳ>2EP`iaFƽGc{8=ܤb =Q  % +0AʯmB'!W_e2Bm"UBRLH3!􈌎AF{BP1gʿAڠRqvs5%[/\.*.ݦp1*_ǻxq=1 \|'x]_0"6ۋɵ{\\6N>M_&M7X8q=o[0v}=krk_F]\O툴NSy5\n<7Wl$G]L%"qi #N; e" HpV%RxIMHRbT&8*M;oEz[:19':k~*ּ\ Z9M(+߼,dBBŬ @1lRH&GLhMO,bP[|4"ƻ@odx6lyx];Sk(v-قoDs36bm|TL:n5vl#3+>;jG:P1M% KUN@`>_iEN^"wf; MYB20j;h("aDXb1$sDG̤04ZX Q$eE3Oh->(J'08.*R !V5/ƃ4IKƅl]d H#B )ysR2e6葻)0b8%?l}(UdV7{(~a~Y`Ikb7"8fMx)#4L7Vb\&rť%&EYqGv9æDЃnEsWV>>W-ԃ5S3*MVEpCqut(m=|RAT%kP )*#Xyr&,D0h$wq}_|u[z y5,FbhmN2xQ !t\a qsȜ<:y%X'Y&B)'tJ aͭ=ռ-V9fɮPgg@bv>:Qv(,5(h+t@z,$O]|WF=@s/U4]]ΚtT{TB(C/c:}ԯ~ՕQuL~f\CJ٨gݐi.X!C"SmI~j4>=9{ߕ, ?ޢV$yhWD{5_e%x rvP8=P0i}WvF[̻7>'#je6XS.C>ǡf3˜; ]ӕNCRfW٘ JXRl K|! S@cHxQ2 '^q9&.Vg˧^DR_ \1TE˺YYI>bvykb4z(z˛s FkGVv=ueڅ8Su%WO[S! MYrAFL.28 `\KϹ"t.X7h*/k^\&a?ٽ{L&<̬9bf6;؁a}G6>NLS\DCʔE-Pبi%%>Vx_Cm" :6;1g9cTT p \)-F͍D#YFz!(:f[vдAMۥi ہR,9$_3Q[~kN͹=!taO R[V?t!EQpv|J̳jcxrs׸(r-ht̳Ʒ7޷yUvwE^3xf|| L[AEPdnP(Plw.yԊ\rT B1\$8[o uB>$xN%_k8[,879==^XnܺIL1Y pa0YJ(4p8 .(B#HB_u:e[އ1K0K!ؑ#+LW=v<ց izRR K1c'P$8BWu6KY8h5 F(-""93NGBiȃAt~ f݈{=ZK6Dh$ 9YESYDtr*;NQ*o ߴV>zY TWEzd2f % x{B"AgR^pRp&a#ԖG6duBd)ᤠC(~Ie[m:e8M{v:)]7GEp[t;>]~W?KSs/5gn4[-~74]/vg H)V5M]9OhOWcpDk`75 @v>v-r>]r]EǿQ[{3Ul-aGjك:ŋAgH{y|\ӞӺ'4_-ĎlIhyLu/%#S>U>;IHgE#_'\34Lf\2xWˍqKed3.8:gGm.o[G_pV3@)GatP&2cqpnuou'TߩXW(͹VE',/ZUMB:4oXܵ=ܒ{g+oqhˋOe'=vdEGtRi_9iEfcO%v\57E7OF?OJ|ɺ-+ڞ.U8$ vc2WH ] "d$Gx& 9*2x,g˂C )s,Xcx !Ր(gZ`HޅA̖sZ6< \YrEl&t;!C>(j2NxƟ9Qtx'%.Aȳ Em4/a ԝ/t=_1 7ؽHn}C.~ ~b.8o/nuwuӷ=M}eGҲ: \plM;/b[r<̦ -nmd~+[9/}RU>;}:e+w!o)_46nMskz /H *ؗ3|pAÿlzǓ^6Jcnv퓟$\e0Is*n gUuه{F#G$C(==m{Gg?inҹauanndc'4;?cZZ\`gGz+xq9= T>ϵ*iawnшqt=-z>~%<=wѻˆ_}ðԟ?^Q06`o q l͜0Y)n_c&8y`S`N7) `k QrAp#d&iP_ڢsB87Am$v>ElQHl!e*ue\zx%-uo,tKeQ|R_N9rv&d*@-jbPܸRƐĆ켳ّgWmܗjnCfVJ7˥UYϼ3ҹdR%!ȈJEI*WAZP%BHU̳qd{%DEBLEDci`(wUΖzvxiT J%ѼKvC&&H)rc}"[+:jl*P2zsd` *e,_ߋ vz_^3U=¤`%C ,ue$'o=%\413c I)?eV#|-F(_5yʷ@B=Q3`(6ؚc명E;IsCT-n^Mu}*5i&ˀwHbjml3'`t%ˡymƪ#(לBd}r&,TR(A;$ vOW5!5yh9keۢ,kҭ.dB[Ġf9:_RW`/*PYͧgz&v4 3:)>R@:F8׈ANFڡh@8 j@zUynJ_!Kk9 B9d,)f^NNn lP Q݋ʁ$ڊt5>8LKY+Y ҧV)i4pX j,}`aTt>J??m3=qVF Y7$8 .ߎkƈv67:Bx,8[eH( k>%(3#=-'G@6$v4rvqz&،oo/s2mqaxͮB]P(!bɘY+T2K/'+\) 1$`\Y sByW C`p5)EvU>+^>%"HoJ:a6Vȫ6-֭/̪n[b:T jgZȑ"̧]`fEh`;  XL0E=lfq[˲^Zw$DQzTB5@$HM 'Q)mN"؅UV_/];J׃^w[S5ysx7?ɂ_W^Lf*V*d h0Lч-"et\DԅXry{=[oHŠ6P0(ec 5lYR,),$"lBze1G 5kTEY "a6ـUbd8+*ocMBքe{I%CJIK!%'=f`Vi Sf_%ߞ?Fٖ\#=ιk\) VS(3a=]஗'uoa0R:L\4:HF!\ C)W62jJ]DLJԬ } H -@.b)(IYjgڇdD®tEΎys>?xf__-v>U#3f3 9`"R8M-2ǐ|&?%ʋn3,$neQ"Z-ڶPGGDdςU؟[%({< Ō{Gy[h^% g'Wd1ʤ2Q}v3:}׎;rD6aU=_G7M/ev$F#D EGd1R.RBu`f ,nY=,CƢP<o/pdLơE_tZ{ Lk=ڤ;viP SgdF9AjX/κk/e0Ҍ.G5u>#~ۯ7wדj#x7qjtyQ2%OpU*kj/?Jk[Ssf )UJ'ۜ'ox|axƶ{~aa\ѿzݻA%̖ؖ;cyGCtT0M[w?jA])|YwTQㇺ ?_1 ~۞r5$}a9Kg8j|6Jƪn"T6׈= F.s=BU'[z!Zȅg@Q @ yv#.-;۠0z6cK,]3}o?=z*Q|9Fa)!{rW[W)VUPy`O>VꑟW5jWE$)|8m_4c_Gh{'18m}DiOwk$=bAF|?1i4falol^ ~o_GBPDݲ*E8Sk~5+BLLB:ܳB DC*d/0N#G^H5+QadɻfHBpz`㕌rlA"T1nbG53F"<'|-.O tI:X?{! s&վj$jЋd# L]a~w<-=}| Q0]A:9g,Z2H1 ?P-8b F sE۹`  Q,2[uFvpϙb%mf̜ys= uvf䒱oխ]f=`}^ !9Iy{CjPjX@A8ac: BuVǐ;T= T!b 5"o&X/o@BNPl l!PJ= Z>%]lx=`^oQ=OGg^l=i| &s}T?v/24)%ՠ4EƖW*(=e/^V) )F6pSv3%b|PNt29O-Fx˦5[ l7 us}Mh7}a煕z~w7_._[њ#E)p&B/m R^o4zr*NWڋ%G/@^ nާ(0fz%9V?䣘{y:3^5y3ٺWmpʴcOʥR crz(,ebJhsm+㐤UR()i,%l9E%GQv!7, )O.3wkۭޭ-^=p=E5k.jy>7^.G.[K %Ϻ5ruʂ Ɖ"oX&Z^~hְ֫ SM6dRWa;:݉vLO=ނߘ?ZRt+[^qId25D:M#-UcWw\$y+1,JtIqjNŞ;75i7Xs56V8_w,(md*-JDG_E_ӓruy}PN E'ގ'F߀\T)U2ek#FAd{!hL1.O e)S{lQnHj0{%}c@H?eS0놡!=E}ph)>kxSYb aL9J0bY?F1"SDsxe^/= ,{Y%0(JBbfd?Iḓ ꅷᝢo$t<6<⒗Kz`{(}ެ:PJْ{28+{N%;p@lHbrZRG0toՕV9!u'*R*{RWWoP]i +&X<u #uZcWWLԫ7HE{BꊴNF]1쩨+]cWWLFѫ7vKQfo?bʜiY߱\]4ZLRWfyi#݃,痲KХfaf Rw4)Iչe.4p-f e3iY=CGîT\W={{9Vkomz|7Fq.I&&Hl(R bЌOdcCqQq,Ev%w_0$& {gszN~Ƨ(5!{:+cd A Iyds/WV*x-t2#hL0u l Jx:QZY7 Z{忀," 1At2gJ{ɑ_]}0.4tAI[kYR+e KN\Y@\E8^@qaw@a2h" :-U*Ujg)#0`)IX>|#w^ŖT9`0|(&!VX;؏ Tg w;k}Gؤj̯㚳ݻz8vaDҍ 91ZojSm4؊9!l*DQⰹu6-6#V.R=J)Ԗ^zUD8gڊ%#KT'&AgB8s mI1W%l*a2FNpo,3sjʔ, t*NNYK+?lvcY8 A١@&QDD=DBTYJx8c:C/:PAF$fD&~L#H a4W hJ5jDtQq5@ ܡ nC|ӠѠ׻,e={^0,p# & shEm].P"*l8_uMrM(<*i` d'e5Ɋdc(=7GDb"U%:k iu%FYe/ZCt a+x':3)LuĖӾwX1ZJlځIkw:)ğ:u쥪nw_V}nG`eGw{t~|NJ$8PV(|OlLő-&(\P3y  2\~"aE"{%lf͚ڣ2mS&[L?SuǝK;SEݝKC!7oo?&1;*~SB[Ԩgz;T}q M]fl8*i$}e .Fտ?CM$[✈r'RPqC F@0J('fHn4۔Í{7=Mj_HtSBb */5m*6O Olpɣ\TרXs`VkQӢ߾m|Yv7 -OL^=ԠOM,nb4>Aܿ-H'=~tCԊ|=jOneqsН.E?Ӣc$ŢKNS.} 5|o YnbhL=^7 }r~ɚ+Bs`{[[-&^zTbT21C7p[*2s"!X(OJ2գaK\M%ʿ/]HE_Rw?{?\Ѡ5 -TR$ Q&wIo!:7qɍ& ņAݨ \eFEXLN>wxڜo8FǷ=F7scּ-Q}QS1N!$ISƄ)C:$*E0) cPTqGH_ŕڵvG6~;.nΜ_k͋?iP^oL OG pg.pA D!uBlOML(JmJ؀B| 6:S*)DEf@98 G\ {!Xc_HǬ(-F)8Iiʃ,eEQEE5,;Msm:aմVZM;]A)dF{1Q{q \KK3-&M~gmř-ԓ­_6$Wr.qO++ dV(\Q"*kXreDog'ͬh3+.n+vl|-G,@'CbRE$>y)h xtQ먩7!)ERnCrn?|Ї0zsvlT 7z.i7DPW-1 &#>+FT!^ lgNgE炵𲅗-Px!$S^P+ EL}Nޙ3GE4 g&GryWQxyL3կ,Ϧ nvGS/6~uhbS_ ]Bۻν>!e[Wغn]MZo_tfno_ͦClYan[w=5=z^h]Ffn'"z>2W֒zN;qpqew'ra[89N6rSehMTL$ӌPI0 wrrd"US AdqWB # ?8"O8"O8"O8"J41DkmKD <琒 %ऊBcݓyO@\(ҹͰ։3e78u ' /.9?~dslѹБ̞,陯m>BȲ*2VLP|=9h䡗,1VPqVOW?c͹` K\/YeF=_iot:˟~'V?OOͯॿ#z^^d>X5sZk<Ⴃ];60GQ!Đ f,_$qpڒRKzCs@Q哵y3JrApn^rPD (n)8G&A) ,85>DxЖ M)93O~dw$}׭r*R=iOduz<]h %!׀((d&}DQe_K w1m E5,﬊ReCL@ 8|GLj͘DPWX gY_@sUOtկd[AbRa9h#Dh)_p2x/`gKđkė^T2±8<#1BrmƑȃ:FA& A A\{1hЏg4S3}`];;ZPy^WM;^o?Yt{V곹(v62뿆 .fzEѲ)]E*Ȫgn7S\#=iw5-vlQO~irZ<^65*fS̕Ⱥ zu{fjJ1;Y^vLDz_UdWkPҰwp y6..)hk.ʭlcD-C:)Q W>[F)?c{;rdGLMyHSڋTyyg&짼sVJ=)uv='q󬋃e,TZpy%I> &YT>ʕ=q88)7Sir3!_xh#H9@lo\)E Bh>LhֱVD6:bn 5;a۪&ߋ+Bιzݦڟ)^1W^ĵr}yB? {1SZrE)Csٵdw[ p3X`oEEHd=Wl=";Vےđ(ǯŪbl"Հ@U0! oļ+ж̷AJ^ƽso'}[)hlHBi D=sQ1 {8e=˾}آdɦt9M)aȣaAh*lGUj_:ܣT=oУF3RWL5lG]Ujb*Q[TW?}ueb Se/@PIH̵~.0 uUƈavmUdu(N'LT6 ZKcn-)%g ,B"m3IZK%&F%^rwK0x Z*1SAd$!EKe/&1\AC.gK4̅֞ly-y/RB#MIm!xH(D+T\N `܄9S / 614!d6b~6$^yMk箄>f{/esǎ xjC]Έ7Q.^pmpѦ`)&-)F@P :-\tɢsRh& ³T^Jz:-R:#gCOYJeerS ŏ*7ab浏Ҵ ߛz->+ ׷vO,"˧'Fq5>}"-gsa,˻Xh ,.hc3+%n*[Z+M 2xٽVkv܉}ƜAٿa$cGWȂdJr ; JNix!I"8Ǯ*(A mSA`A"LiX:#gO9k귯@̖U0jyHP| K}WsCV MVfJdI 3h$"ХBQ@R #\ Q6Y+,e-0{ւW fjulI.:'.K`/CE=s)x ggDQY@ YpG'㤊u/*6:KtYln5Ax7L+J";^ V[(Ӏ"-uf*!7"xXfnBMn==J]KOv,|r!|{;jB2h]1F$ L(=??iٻGcF(KX&oOE::㐗JX:ԷVÚr3ˉxHYUXD}kDZ}cL]n,Z&uG=NqT<9gStgsfʿ!}I哩*#ckvŞ{<{_2^)ؘ ѦپF]72ֻi92v,qiyvtrJ@HIq֢Yg%3db.E`lF{N(${ iT s# /|zOUsmB'] ?͋>ci{hixU祈͛8yNtwNV'~cv'5lQf m҃ϼ^?_ dw\Pkhp*#-|N{As#yFv|Ú|]rd A#JČ^x2Ȣ#9P[g+^TfotT{Tw {g_ITOÓTkftWm7=R=!?/R-9r5ÒC̾QOI0&[{[/'=<殷D'opY㛸&,rr>YCHWI컙-Q3V,<jﶻ1iʰ2 DR)58^8}0`ڐe`j,ً ]mwE~=?)%jo:rö<1ܱ}VUh3fHRT^ J):șI9Lʂ^(V d;d!  MMr=B3-$Ɠ%mTor\q^f@&ytR: ,d{=W[bƽͫ԰l2OKX HIe4D/-6r>{0Vp6 [vS:_L>h$ARQ$id1 SF$hk̢kR ߬5!Eԭ (s1d,>0!j/z%!i%-ELꕸ31{+>*E- mx#cnk n㴆nF? 4,zT]G}xZkuNogy9nׅkW?[y οj9Do ݷfoz,Z^Czy>E}x/ .ؖrl 7h$paȴV QPRH-ŭvlK.tJb:cK񂆾YJQ$@h\i}&7nݤsUE@-In̋:osfk$=jMFЋ"-KE=/wco*k&Qoo̙xzN3~WψJ,#~g^FԎiв$t(٧S~^#YN9gߌ"|v3"&h`rSn 1V_c?bgf=y[5E859QG) F(G:b옓TE]dJ ^o%}{tө9N%גNFOz(j},֐EM2F2H9P A'YvUMӑDߍe *DYnF#M L;GIT;v5LOLcg|CI6?C{~h6o6damw'gm(xcD~sY1qL5<.Yɞx\Nkl۩fJ%FQ Z\Tz(aBrQe,lAJ"+5cg֌Nl?^U(6;ͶT4͆͋?Єpix3}Z΄QiU.Yc {/}Q -Hc(iׄ#b )eBnFX2d)L-+{b.&^/B~2ɘG)up huRgTA9&dӄ)1xUt`j.dlC^zI{MI;k {Xnx?VEUCG_m%ϴ${d$GJ'cw;M'ɣe!J^:롃R@㽌 $UdEsGܑS{|}Rf,+((IJRPi҅bR,9r2b}Rq'IŹZu }BY/>d Xts-_-ݨz^KC#ALrOh$[&P7hոBTRG = !Q#H kJdS#sHŒ/SI Nge~ Ӏ6YAzag"6TinCBۧ~?TR!nxtut3}z^tǤ1;ʓN#<m̟|QYbFnxe+}~}x~,⹘|9yqrsM/מzKk.9tb{OOcc?wn=QNZϭ͏;RyBc0}jurT`4> 6404т4]jb޺y**Z"ł`#s ;^Ec (ɒ֩_'cXBZ@LnJ'TS(-g|i6IfqM Ok+;cGJf?W]t /Zri>/?HOooo޷j٪};p89ƶGZGBZޣl哯6N~%vЬUhoMV:n kՏIF4!?{؍؇] ')`n0,K$w{}GG,.r n[UŪb]PL%B2],)҇GE4{Ѳj! uK#j֜I-$gJZfAQ)b%+pB8^W5W{/,{Gˠ,yuěwyB>6!$yRGg43:`)e ct ;-¼N\5uil\ɯW o.cPb*fh= G\/DGQ% HZIcbR3VxW$ Xέ,.Fy˘+\2&Zg$>΍$Iōr&lQ xPJpvDL &mSʌbVhG (O g #?rWh$`}4}i@}aV܇{:fA;6wG]eFs؉d+D@XVc$Fln/qn-3;Aa|_wwX2k6{bPЏh4781˗\W&{2޼zZ]wF>|:<٣Ы=T@{жFYet|s|s`}RWldA3K]2F=Pfx[=[Z8b{}r B9ֿ;\cvy9\;]Wf)/'i Izz8vQ ^GjisdӶ< z;5I%x'+ĺNY+V; mGV&x.ַ^Âj˪ ͹BKBNzze]&31mQfPړiKfPFkg!JNL38۔ȈWn{3Z]\eRt+!TmҮ2`cZ#"qѲW%'J)H\eZF\e"$W/2JJ;qŕlʀy{ WҶV]*ԝ#+М3"q+q UFW-W- +1Zs!cp,( ~$4W5`s2⪁k!hOvqjP^v{+ա[O2P-W8q2hhvqQ*щ($>Uf)q^վZ`.%\1؉+.\8O'~wNmAp? Aj_㸱pgq]>{Y Q`gǷ󇽏XIBxQMFOsԂ3 3]ïK26Yn'?&w1O/dU"US9Ld tw!RԋIg0EZЎdKfN&Ch)* +dX6$!;˝$Rq]%Q!U9צY60gp[u[45JY4Т`@ ش_p k&UF)뭏($2mBEUU[UF+UF[W)E*hpnBET\eӮޏ2;no{fE˹ $ 6qeWWo=(Ez&d_ ߆ç~}n>5';dmJSuY;?ۉwuQfNlۯOk\ DN;-[rHKG/D Rj@4ֲ5FE{@ԺԔt(ty8ݬ>>(SQ}~ tZlW/&ϧi8\s/NRH]DA%V*d-ts/H#.휤:g^lPo_5Z& ݼ_|ׯy!o-/EK/Zاs\ byI<٪A  +!2+N' 7ʙEG&A) Y1q .DxІ M)3Y)<*)?[wj䬯ǴdÓ{sa+z5ߣ}Ow72:=4AkI.Bb[d^.J¢H2ʀ$_Cz+P8kT4*c,\XFz $C RTk$L<9Rz|<'iMZ_Ȃ6uWɑoƦ- vEgY^ȲT-nY~sKNgG~P;R\\eڛ<곿7) }O]Kznfo^B|>}nnjA_Mg`gV-aU凛~Э_ӄE\ԮE\4֪g͗c4&qM䧇^mr7D;Y9^9(x :*d{M>;Yf̚0|7_ƨ플Xnl5k4kUI|Yg̎1;Ƽc.Y 3VgfPohw=J;(j#s{,tږqAo7?I'I-;IoՓ昂jyiIY$'14ԌqE'ƔŒ O 9>7 jga9_9ݓ`cX^ <1\{U6,P3j;yϓLcn̘RS2q=E>ۆ/SM;eU:(tסbB4vK8k@i5nNǼ'|/OJ&?}pQH9%1ZKNW*)o\)uE3|.TSjMπp>n&8._DvxsӸȵ. Xӽa0:Xo.<19ʣGrnk5ݷ-AJz{vօ4-ɌjSg @kz;0 vh%g]oہI#O6 B5֣hmbZC%P ;vjϒ^swe5?g¯X֛LhjG ~P0"P^ǐ :@yr\ mIDYQ>N{Z AJ%Tr "ѸG( IrAiNK1,gw?Ыϥɑ[<_z g=_2 C^y᣸4xjul>}Y,k}^(v)?ꄲZrL#L@ Wk睤 A#źddӓ q8ȝ\ghnLVq.q3 BH'#WJYZ) #j=6z$YG' 9R_3 "jЄ+s#A AX^ϊّjlܼPsxiŠdi7,O5Xz *-xh JBAN_EHE(8`k(u%x=paxi#_ tV1UOIsjy,JPU}ޅ`s{k!hIDE 8L "r^A9/(Hɜ|G5]i)4VRS0rؘ+XNy@M8sH2 zD(V97z|5 gHP棅6̷{AhNRJK<J9= :I:-${̕] wC|zۿmJTyu߯itm\HhY@Bw8 e!qPgL_!9H'DB.<)2E*R*ቩtJL2s:B2Vt)fIרR Ri2K󂁞Mfi\s.yzfiJg~\`/xs7ͺxT0=4lᬘּK\cI'LoS¨]Q #eH5iX*z\⚀fyx8m9dͅsEm $ r7 axZĜSPBJW.R蔪0;758&C{(HHW+R;9p^eվv3ja]~jXhJ-^.l>!/g5Ew8">Gf'ý[ EA\'X\>;{P^c- ~6JRx@T6F)D;MM.x뜡 Dy &g F8RKھgX$E9-h+D<b\pft"8ZYדX*wהKhO63wlVo~;~ if#9Yh\*/I Jhmc"^6ْ ^|K%!.Csa4< )D%L{(Xģ%8籞Ǿ!3" 85I*ʃ2#༆(G=q[D>5/zN9}rYsb?$c=کɿN8nGvG!*9N8 x8̌3;TS1hxQ@㲹;Sl}{xUmq uȴt9.)R㥈,QJ\jkLpҔXiSQLٸn"B3H#ଶq. Pۻ MZΎ-s:? 1\XhQN'ZxgH"G3D"Ɠ03=@'3Ԓ-</@x:Gb, Yt2L\L!Hf JJ7]V*h"sړ3+Nbqdžp(yoMGx8<-f=zɝȟ{ʹ޳,GT.eQx(kQS$2mۓ~ȿ~^5J5HF|vOMk0IId ƘJpߤ[Tz2&cx["ZP@ǃFS!,92:|^>Co;ߨ+Pos Z M5[_κqz]OL~\VJON6y~[tqQ:?5%KR\I%oK|lRԧw3df˕$nnWxycp {~ccwo ixI 6a͌[QjUfZSٳ}Pq:m'eRutɃPntKoBlJ4gQrr E]U61| ?=YV1+xj9Otouj(0J('\\+4)>^SԤqh췖$Z@qzB1Go qrzZnxħu,pʣelnmX,oef;hfDC,g[dM:Au~ ReC5R}cYxӔn~ΞtMw,nrEe5#>0$a.Ry[e޼uO:/u.ɋ=h1 E85 L%k+B 4(N&y啡 {ԢV MY[]Z'dōf93sSXVJH])h1\Y+HEpܠFB)އ\Ƈ|R]Ɗ7e-t9~bMЇK'\RKgQ}<̆4P HBz%%PC]ި &N3'rBR!W| 9<xbG'h6@ q JE:\r!*.J*",v4S/V7C4|*o,d >zy$@)uz?eQ2UWlR)]Q8^𲇗' /3$; w歕"8DPD&dIW\^Lڑl,<æ5sfܩ-SwZx=-rjug3%x!< VȨ/[x_L&'BJDOn@o3uB:QkoNn|]uO'|-׆ut* EWuVJI✻ B1Ʀ{{dtW[[}V[[o& 3?~hN!?GlOV] d i%~߷\?[-]Uzn+rq\YEF=[=5>SxqnQùh#=5|ʎq;t |~L'2Aj&VbewKBk->O9Ք(%.IgZ }{c'DOHgR5w0ǃd8.Dx@ L&ʌdXGh)l[Ύ)=g}o$_ ˻C^Cσ|rv̂1{p'V&nywQ"IHLk(Po<2 igRiEL@᭶&H-U1r"S%Ola.g}|zҺҐsg_}Bɦ6EaޮaGrš|a>6w7H# W/T'[Xi"VH>ގNH8e;z VuK8a󗺀~>0u1Eΰ]ɧ&7qκ|7nͼlfMcrR}'enʡ;fA|Iނ^ge\/-C8ޭ'Nzp>W˧t8LUϗ3{xdY.X=S)6yi I諾JJ-U*4lq8" i2.t7x+2KSrK.^ҽ!mRơ|6F*%tcG.F;ưnԕK]A;UޡH|2vcnRdҕqhw,<մS*-(뼈G_oE_mfųK8k@93UoǼ'|_v mƸ,DsC)DheyReIR&i,ŁD -ߺ}BkZi7~V)N^V76/Y 9eujP{!$!哸BP%Vqڊ)+VָTC5i+AO9wbSI^~Zg@zf`FSk.32F#!;NĀIKo%NHA:(!(+ 3rK>H+$䤎GvØb E\'T٘Ŗffz2ɗsiζ_nR\ۃ;/WSv뙶0Wsc_bv=->)e8h[D8gڊ JKT'&nqeC+&leb*?6܅(4L2t18q{c89J)S>jKS$Cq"gê3s>?r.Y=K|0H @35Ғ L ,%ziAw_y  PNhQf?-E9h^\8%ڪ@[ig*PAF_ B5=ִ ! "0O4`B̛㖡IJJ&+edHM)#"V5Bą:ގHbVQVhAYR:FE*w:3)7 u]}Re-ynw Gϋ1S9 }Yn"dXM+O-K~?\+jpX|*kr~Q&h5Aᄚu7z&:1r=_C_3HPT^<*m֬o{T\Ocb}@9A &K[ F.MޅOQV? wuq'܅Xf绊]%?E:LO.*1|=SV2Sޥw&5㜈0dc E(0J('sA:gmCǍpWwm9}6(NzJz!ԁhKUx N׬p*Σ|*S~UQբb-EJRp8o40BI}[P\_{ 阫SZiIp.H= =w(Hh]")?"Laepenne.יͯdEgV±]Β1oD|RhgF)B}A[Ѹ]T3P o&Kn1Y `J9Yޔ(Q |tG%*#6׎^znqGҠ)bև܌r 4O$6䣿yk9 xp [l?]L2k.Jf /#γ$U ^6txS;yN wv+0)L%FFkeIiBre!q X?_vxS/SX#vI#¶6 HJ f$PO1&$ƿ5)BxohtH+>-&u%1&XX7<B#E/t}]cY߿߂7~ q/e%\R!ϟ% +m>boV:Xor=|uq*Y _@lNu4o1=4{;xu68€#ǻb zAyoȹDy%,wVzx7Q2uuA1/kIŏB2r|wJh*z(FSM,h ,uSBl %ͿEp,£h%7mF/ךа9$e<UJrGy_uJl[SHdtU/9:a6''U,KꞪrk/).ce B797TDU1P QcBC9ՌcZʘ8ez `zMJi*(h"jZ36Fz͸i|5x J{؍RM33βFJ%Q# -1blb‚*H$hӄMvrX&jB’Tk90 4ŀ%@%$# hAv }X+cOHƬ(-Zڤ4Q @Q@ Րs6!cKZIk%}JQKb}Ië+Q?NxQX * yp)!Z|GZY@ +JDi-V#_Q |=mKr ᶜ}nlCKAP :QH!H %t`)h$xt5hMmPt#A1+>E xyϤJr֣Nd:[ࣅ19:xbw2K-] pT{.X /[xB=\h)PDqɞ98*D28hr$_k^-ɹWS߇Ϫ 욮]7&r%iaUe '3 7_%Ɵ[utk:YWf!,nnWMW=_7^ =/\pESE]y(JxN;_qtp^gN0Y9ٷ۔7<}\Xr*TYnmHfL,#Ns"NOCO?8#؍\*p~~j7j/j;QX>&(:g0G[ b7ș;1^Q'ݓYzAu|߀zσ.|mP<ݫgptì|\bZ*;N\^0[bv&V]|A7 u:EqΦ|+5ȍr4'b4zyTL?KQW\cE]!CWWJ*[u +5ԕ dѨLǢ2T]}J]!Z٣QW\KE]!CWWJAZuTFK/H.յ!55 χy:#wnPZΈص>ϡTR+/OK(!GyR|n~B&wƽjz>F. 7XVs 6ٳKJ2,Ԃ:(}EڛWݍ?+NvYC`ܹ?[rQt9 DNLQKodüS.v<)N}U(B.Ola7y[5˒fO_uw[+KVK}֯>pּ / *XH^Mb]$ W6*ϊm*>CbONt:0w6`PdιgU {GͲ 3sT/ؤ Tu2=d? v<)j&Y2/MDjJD"xeqOdKH-.3 :z_/0yw1;Z|C&S1τ|#畹zJ-%6$ Т11+?y[Zҫ.Wn2:e/Qh,5߮'J)YpYU>˅+!_.Dr~( eM׆lZ \JYW[&|RzXX!bgB96_*ωUʡ:#+IР:Ic`^1!$cɒfcq"AzDV$BLD1ȐX̡gզӑZYKw./ua蝊†b GJ'xL)KjB3AANK]͎̣uBi1rU@S:{,y vIjD=ԪdI`nɝّ̀g" >FL1fY$21s"1*4~F5~#j6~[*wQӂ@FgP*3 k ]$L4SfIj2+zDhnD|+5Φb59xe3ҋk?IYk<,#r ?Iɺ:z5QlMX\BQ;ď}`s=?ioY|ǥk@T$UTSl`(M[7q++~tWSBB$|CZ4Jؠժ N(͆g5&/͌uJ~R^.nX8Gy@{7* Ŧjq~rBJ.uT}?f߭8BJh )ȴ2DL ̴ 3֮x{FM=]I&['3]Ҧ|XcJ@r6d qLs>]}\}ͱO=]}6usEyHP-z*+ʮ>Sm̞ݥRm; '⦋ ɷ֩?#ɍy{><{!t%u=u)H`&ʨ,Iz*&霹VYFPȸqզsM9ZZI4 mOphǛ4M/jэ/bjOa./.6}CM~n~~=7딹ߚC(xY~舋6-g0󠠈cˎ-2y[?ͧxU{<>-HziJ*ΆVe$ŗkWzTpr#D{q~]TwNpab,}Jo#FÀ^P!=1 n>gppnjI.]=: WTsJ-TM+`„ IӴ%:T 7_7YE "g6O)Q]?s?=%2 k6|,-=#8un +>\ܠwm(<:zߓO͓ߦ.6dWQW[ˡ+^2u5N{l)F_]whv\2yBJ uɁ,d3.gזm.åW?~9&z̮ÒS$r@]`,ym`҆4m}FQDR7o\Mu--*LsZm_[fOq 6|M3*6 SQ~;wd_O{Ozh%ТDW#QpLsCm )YF3oM.6D{Fwpf+{Y:"&Od6; _b"h%O6X䖌{Ңh,>c6n +p3/FHg+R̈C&H< :N+˽I#rݶlʄn{rYF!w_de%bMЇK&\ q4a4e\.{i2v-,Hq?N8^zƍ'C}S!+ZJ=FF9hX |]8!zUW찀%|VLg Ʀ4"F ~um92 Өtr)!)nF) s LfQٖ6ݶܪ:Ă6.i›1Qdk<W4c#֏?3}[+m-Xj:٫ckO.=9F ?|ci~{0wOA#btG\>"Rdf_v}n+gs ۹[=9]MɥdtǔW\UYh39rCP{S!dnwM8RdM=Yz"|NZf$j yDD|Tk2%yU4)fQPn)23H][LN7VH-F:zkʆithp=jrGEjPN:x$k<.5{ !e _vhrҺ{7^ G*>{ܫJoD&PfrJ,6`eer'!'5N٪р1fY)gb\Ih3zA!gl8d-Mgd܎RNe* ,4,<,\̨mP|zUN/v^d0޼7=nǓ.CYs*Vl jf &8ɱFWJY6EB"645oTl&sf0)49̂ڇ,[ba\2.EjWC6VKUY hm\d$bb$ to0c֑wuꅳ]zy9!pƌUd%$l,4L.I[ 2ґmJA<紞zN;&k;S,[׏Ef$}mETBgslH`!fV:QR_<W ) :4U$4hKtAx>2L:ر"mKuIX(R,0g)u$sp#C}'kbC֫Ǣy%Ŭ|Nq[䭱!s9̻Q'KCAL(̑ῙVYI41F(aqYDK%zx^),i鹒J8*gt0'S^AKեUo) v-~Xy~[Or*Znm~"\ ~hmP ד mu:)tRL7>kQZNVtUV{[WoVG8^ H0k-8M֒%9x)CX-Ny=k>QʱGG'BGtw9:ZjW|U1 4iQW+ Y̬RF94bnCt9Ƀ=2:m}oe.:J=^QC}b=KNҮKO[_nS\ʃx.J"zP)4}Ec>F(-ao?RpC*ܴf k.LΧMc0+yA_KΫ{ہ3RHDH`c6F+$`l1 K%Ա[n((UEA;]*hKﺜKS2=W ,huU7*KbT6eۆ3@ h*-seFsR Rd pMgdov? m8~k:w߆-XCa/nޣ~$.Ö$ysϲ<3`u~hU {¼3A\dzմ Uoq r%?C̛5oNd08BMxnHG8}Y$f4hc`54)5҅cswB(+gɀ2 h)CTR FX#xW? %( AR8M,9w%Xut/ԡUfS}AV,%9 B Kf3eE =["RS)&ZH0<)NgƏ#?U$`~O_"o Vh=j*uyYA2>_Ԟc9e>W98"a`TE X(}5fi."6;y-Xhm:ˆ;6` RA,9,;q;&[B^ݭ|}ڻ}yv_Q<+yZ#/t"X ȕ^B'jWzh_hcTBItJȸ6) ec/M%ÖUFS1B!2[BpєcznIҠh0O7d/6Z̶LjAMfO7B@_>E0ٳ!W]]Ϡo?k^7үO>@ksc Yߢ"7EnnRnUE@6>^Z$)0ԡ21M{Z`>)\&%E2wHf" ϺHh\=\YP\X(^Llaja+IQo2as&pQϱ;md=\h-wKZnh n8_ybWW/ tY߳0S§3ИG\w`QĉfIj%/j,Pt/'163XMR<SwBcȁN :}_#.PԴ1U*wXv;<<˞b6ݼn۟+)t ʩDlj2R=EuEojގ~_3TT3O"L3a   Ƙcl2KZ$/ ,n.>T[Ƣc\Wx|Íucf"EH`0/6~$0Ox 5 FwVOL~pı$Q+-9R 3MXM>&W_l,C ͱG.WX1te ΜXLjAo};JN}7É\*%y8Qn>y8QZ*$abu%^UVܼJT*^TW#M+1BnU"~U[WWJYGTW2yU"ޏ3^UV[WWJ)mF߾Fޟ̸_Mߕ:Ft60xiGh3?K sRk ٨Eٸ0b/_?}̸'>Xʥ;MfP/ :+oT+'݀[Й2Ia/LJ9>rȅ~^Bd40K.?U_*E*VW{Z1"te"% SK)JҀV0XTN-fIj-DX+W6:Y&C@O  k@k5f,`ZFL&Z i虍=4_QmMIe4~8mk9]ZKt IMՁ9SgcfLXSƳz&k r/iD]56)jp5dS@EfX]Gg 얪ֺ[ivu}˻5WFY[\=G}h}ijM |K^hb/n7;kbdrmUiqsPwXJ u٥@쾹NeFXj fs0Jkl+洬WR*b7Sw2*N9DŽRn&x92J#"( -J)P܉9쟃4;i:õKKJxyx l=ӽ > ļTab.6e٘{+\Rq3O* J*0^`*+)$.?c$dvX9*I)@+9EE,qtRf(XǹCpvY!V1F: lpWjRz\TknlfE7 !,r:tW..g)Ǣsg=l(m]ǂ4[M͍f>:Ldt 2j?7*NFZ~FxP)]=4;@kJm%֪ga*wZ|nn?`6#WB~{ % ,uiOak-KzЉmH/c#E=rHUuMɗ?x64DXT 1(R#O?uA)+養KLn"Y9{'Oߛp4jXr4Z~bNwrD|#T(6̳qO`#xXx|>>Дd}t)np0sHW[]E‰uXTWշ 繥{6A0ɵ r B($<3@px^1Tk%izʩۤ &:}S:k!3]֮JfV t] IL(BeBŹ`1X59: Wϱ] WZrV9AfxAzy) x.ÍN9Du|VZl4wŔHu%H$9Cռ!0IJm12{ˌXnt Godwc䏈J#iiGx@֫! ;|8}$ W腥ii
EJ'0zD+T4hiB$ Lc,-gY6r?M:fJ-ɉ#P l"pAf敏XbZsH6etԋ^(D]Z&>:,,ZHG,X{!K ox k6ZE qHn6Sj|%_A)3K]xGb[X%(DSħx JBAFX1 /h hDs8C*EFrrV@1{5e`]@fEm D:&;,D5N8 J:Cpt)04+Y";+? d4=%LE[`4IЎ@9 `A-7?IKxlhT:T.0g7`v{Q$AyL:b<,őEqT{68*t$|w1) /ԀRxcDoaVMT(LsS23m/P\+bcg>_GM&2r2al^cp|Ž5bsUn4~+ґCY:G<}&[2haHUYy{,@"E,JuM4"&B&f @$"~0tUT41_ @FP+] @]LJABq~D,#Nr ĴQMrVK@: M)pZ()K)µn8z<ˍsӱÒoLJmYѼ}ӁC",Dڡg{8W!Wzucwd3f`A NcTHJ2ᆵIռ6`Yf{BxD'(`CH7uB437!y xoBؑra@Fel>Wh$2mt=}a$D#{IxCV"Vc )I ORQ8oR -ቶ,#Tcx["ђ4\q?4gdt2A=}T-ꊯÂ؅VBpک@߆VA|(QN_uFQ]OL_VJŸ>ލyfI]ZK?A)$hʻ~ձqe]s3@JW\;<0~ {`WAˏ¼^ۋaXQjU s*s}Pq|YØ0|jCW՟ +9D|F+d(FK\eXR҂_^eW [tَ3-a~ep wZ _(޺h"LQ_%xJ8l$F:t{k+r0q8A|/7:|a;m^FY™p5|L_<&SZ xG j+ (4Q BZh >fowac5nL0 FO7D꒢Y)4` R $eY, RԈ 4":!m}O*ܰ1L w$}hC %6A.qp dE3 Fdg7D#v/J:o*]O)Qi_%tj) Z2oJE|#U=HpqK4eR"\8-YXk6+eMUd[\{n3wY H)@m˹_=҆]/[ܡn׫)p&׵ݴSSs[{鞾D2>?w=s --\4lA(cV tb^v5*,l(^_4S9GZ[ g-HŸNzikuᏻU%^dU\F'"YC\-1$U\0ibbiцq8~!礱0!)+QH©8ē\xJ<80Zks R4;E)t2^)1^Hkn^vL/9&Xύu<' ,WgHД?%RM|_韌Mao >_xS|ʓC޾?|p+`k8@*ʿd87Z-逴N_J)*]: g^[W z\UW5l +ԿET].Gזqh=tw[9T<ݖ&9x^6R6鹮:?V~[:;WL^jjMJ^ :_}}d~_UJY8`h 5vRÉG?{Wau`nxn{xM+lW9*xkVdV܅$ >!D (IL! c&s:e- s [0xRT6F)D; U\9W!I(Xba8 b3A6 @CQ7fDA"S$Ho!ZGoeiA:RDݘ-g"7ҕU0U;|$[],oN =7ZՀzhKlT0;o&nHln3HƨcWmڦ-mBT%!.wAAp4p= @K1]ɥĺD ݨmG|.Z18BQp&I,I\p^(Ţ&h[xǸԼs=Nj1R@-Dp7t42L5B%r+$Rxƻ"i2v9$VY'X$ >ࡣbSNcyxRbԅ6FaK4 V/[2'ZhzߩK^ڙMz*gO" ^hmί'w{UwFTyGBnT%uԅf$>e.2? w/~ᄐjGdAiptK }oV.R*QKrT7)N NͧWu}YFULs&'~ۭ2tf޾kyX! DG` nL3920N%DtPGjj l^6KN[msr;KKW\3<`A`\yj%?u̓WIy_l C(BT&F ">i{NЦ11N;]|54^Qθb-nKD9BG-v.yk<$KriFCs{( N#RΙ#j5aZ o?usa)Kw6OgG Svï.W ]>]TOmt:fCծ[8]f~yXb晑a8oru3oC[eHͱg=Elh~1 Oϔ4'?m۠b~!OJ;TYWZnS)bӸmʡ{r7`KL1 *lмi$(G\kyk}ΜI \pN<l,Eg%P4UpQjT+llVBV: QكRV-<D~͛-<Ĕ ]Bqj}2eE<j]CU8LCj#F!8XCj!w!J)m t@h`$ wLyBxQo"8kp!&:N7 >ipBit%ŦEbjÛ1A8q8gjj8vU;LxfrlёБbwi(tuI~ Po 8.>C}!21Wpk$b J)xʫj-2}%Ѿkl9P(:X=(&%[ % 8Hgo=U0 6We$GDh0h!Dn1Fs-R+㞵3qOSgW!r%ϡ9f m#7ڼ'5G</|^naN4&(5c;&4 YZ8mgyiR ]2&Z'6,Z |KF'IKf3)XiŏSǂiicb]6 h5ZN,Dz-B80~ll9[a? OܗwC]>3I8òj :+?{ў4ibcFH7{^B>f7TԞp=?ֆ. ڪᘰ3樔QE&u{qb)^L;Be:i׮k6\x gm(gF!Iѿ_?޵&e^UMD?{WFn?͏!94`}q׷w5z- t(OpW?bٲE[Rb,93)9ǍO:hM>pIș[ȹ_(]et)O W]U^D7nbwz0䱏mŪ٩Hv|'ߊJ}Rr!Ht* b}U:.i0b R28˓6BLdR`JZFSK j΁]^ 02%Bhߕn^g S-Q= Z_3WtMg[ I,#>a`jCAR˽rr覿\d3:q}0g6_`{ WG[2?O 1.ؘ&c}p`]M\Yn|f\a۱J]TDYV3Jci:@?)$"b=*eO0 EQ娽4`KU\T)aƛBԮx5&) e aD˹ ]z8 ݙJK4p#/g$_9^= ESWղz <Λ7E+._"w_8)A;c\{Y0^edɪ^qc@4>[g1[\V{l͂lёFRoREːv1P^%v $C4<ݞƱ qR a"1h&ue`1%$XYYΪv)(jY0 ppw +b ܠBdD]APρ1 jE@NoY!q3QxR:F=4@QH6("~AJSjT+̆}IK;Ɍ6;B&* b4e0Cd8@ꁣ pJj::YJ63uܴ j*8R६3X*䵙h̺:$B4SfI*eV!.&)KCi;LjEۊ8N+jýҒG )L춟C3St+I7|LާFz$qGLЧKl']%*l|VqCs I(W4Q6oa~z7ۤfOF5;6~7^(`*;tpj+6)ڦ,DDD@/{Yy\(6\,+a[ Yl"b{C&sS09B;Chj}fR;dzf3{Pn'S-QK̇W>/|j\켯Xvp`rF:n, d`g?ϑ{sשK%)g0.?O`A:^-5xd3>NM0]wd{|MJ.%@M)z༷J1b&s#w;+U^56D}zPzp#(!B)YdNZ",(lѧ%qfrG^5 MM6dن2Jx$"E^Eo P9 &0O'm(B.nrSk1S7T;h6l߯x1Ѣw8M:9kX:"Rt+}_t)=j2"i5"i&&m4,xID *2j49=+IjĘ%=)`@eb^EIQs-@XM~u4XPXX<('5=4p7?ϦC4ѲA@ T3^J (p,Ale(^HQ`SjCy i2=6 !e ~RKۏGZp+]M;ڶ2j{{ktzuJ3 $472a 0Q$Pc~˶-g0F+ʁ_Ngc QkOWu[L_f_Eֽms'ɶH& L-[+fΦGڱ 9RTD!5&Ų%xuccٔ9GXݧ=OX/V *ޔIsT[섍kH'#W̽]rKځV wxd×eB,V^k}'%buK[Un_[Moғng[b҈8PFWit %5-$HsҎwA䠫o]N99xMG~t_.2uE)qDa,t~M/S_N^v˫o%#gkavWupzﭳiX3ɋY`&c4ܲ6&/Ax+m$GEاYfv3X>uc^1tyK%y, K.K)[vev$q|GPAcN;dzrݰvZ㱃/@x-ӻƉd4`}W_3<:SU>d-7Q1`'(FIJ mRb0 ! /b, rtL&WD,ϨoxJ9[kgn2renk5VGҵ-K0k)Zv-S7s5eQFk&7(B+PQe/{xyBHUThUd9F:er]$|Dm\)ofwTRwTW{G)U>JGR!Ȏ>ftzQuu:r>m~F>GJWw?Abi|pAB jtOO;M.5[4:P)'%K@넍Ζۉ2v|^y;F6D5{zFB8?Y{6k& [ 4@EI( v6YW S`jUTϼ0 }l7sr=/ rf 훷oVt>qcgAϷwX88lxjXjRE,mATHӧSTND+)ΫIZ3eTV(AIg7їL?JQecPYA* \+d72$C81~l9{SƐ6ArYs_^-Pc;՘py:f jY>kHZ+E[f^"i! IRgTBI$]zb<] ސ0M)K$B VJr%0S$BrvsZfT P[y%)=޲|#Rj}T|7>uNd^ܓ}*d0G)K+G>u{S͸w{n/HB1B%S/N"*-Ï""dSƐ"фMF$T"[Pd,>{WleIYE/3DrszOMԧ{_ j Nܻ<|" ʶ8'`%N,(FgsѺV&Ix3{QCB+2KKF@R,K v B]@1&F X#OFTҰՓ 3Lp\x\R0#3W 9*YE@BHKJa?H1|ѳ3f]xv_p`hע^kȎRB2l)Y.Əm܋B8'W#N-`L$p$7ɉT`&LO\.Caut-CN9lM ܮxI||ͨH'gJ tHRi,C#Aې]Qhljɹ=fG޽Pieף뻟-/U2J'P A6`4"I)_)Rz׉ԫU3#}ˇ*R&V+3l9{YW|6?B5xi X| M9'.d)s*dr)H(IcEcӀ= t> r –eD>DL S$t _0P3I =( 8_K!*DbΡ=l(\V)HGZ^pt"8*38(e2g1Qx7ޫ CS /Ra-$[cIYp Tf*:FoJF:="[jj)لuE"2t|=؃ 5[3h- R,*{E8'jG,;cZ(KX&o wtF!r;TŚr|"ҴC|gx\1F52uǣQR es+# >hp:|`I[}7ue4~m^uZWrŽ, UvsѲwӶ~#r[l.}Y'B )G|ߧ/jeϛϻpZI/)h>;{hsJA=A#CGzMiFf|͒6PA A9$ @Q,J~H/a*d#L2D{Tw${Bgu5>axvjI.ToE9Sjj%veod_7>Ұ&V>>}0}k59p)u&:ydwu?D0Nd4Hl bXx/F&0ܒ01)fa@JZ>8T1R S-HP~rls Ԅ6f|uᶬY(* 3&(J/ILڋ$K "]Db1$/:Ѥ!4SD݆}Q& $NZLX&38:1(]9],R~l F4-p !lc/WISYYE2;"aI5k$0eg8UO!ZĤWMa$F# "K*d1R.Rtdf),gnAː(>;-359"ozZ Co5%n}zA 8uS A[o_'՝u=T4а4Quv=*}ק#q7w:5wtyZi=5O,nUր3O0ʃ/]tfꛬE+4doOH/Wx1`]ǽ8N 7։A%l,͇3ƹǣp!5ATQrF/Wo{l)}ȟF_>A۵s=Bu'F&`^cz8Jȋqqe.dB_]oɑWtx~.$ ղ`Ԓ%~3$EqE&e<Ȟz|UuՌtC5хlaN>>i#g޸})pIǓT~./p|}A%OБ/יƓlߌI?LԠmk?hh4FioljD3vI8}rg8K<_-990ɫ9-# ݧKv8C7iyXzsh깲zzHϺ:öw7?fk4 FgxE@ `mm+O3/Gq%J>nkǷc9{YyI/|N zD2 1 "Cp KRr& | TХHnZ{pwm]d~ZI&_B=Ǻ<6kJdS>AA%MC 5`#$5CɌx"*[A+(%MnZ:KV_?Yl,H'5[t!=_JZUޢ-IN1e@&kӌsšhpɰls,'e^Q[S!RO[NV]y,A ($#U2Var"$㹶V{g] #mݳljӛi 'r:-vL&9Ct9r*k3A[s(@v,K#9irt蒯eY4'D\b664e2LpYt(3^5-vJpy,VtY2 00(rSҹe\'!,3ThtUaVJ= (: 4dk9r5G,;B)"h|v Ն=lgL.x,|<"JV"J["0V+ [QI.k"Qhc K#]rv&dNLPٲ`IMAqnHbCvHӳ1[MsB5ҷ4X~$*wF8B@Qi"/EUUP(>!pKh Fy0\3 AF3gH #4RX 杨gՆ_H[NKն( ^&nbxJl>!TdrdҖЫz-leJK0.R$EQ(.*=^lI}VXVٺ3NP~(cHK.;eJF&= 1┚b8Zs5F5ox7 rx)#4ym)s>'o B413c SeV̈J}nE;j"\ʷnv%`QYmR,Z$@ Y$,3Q`EwJz[S=`*|Tj{L^O']7oi/ioB $ dAH&"8PBYxgϋ%Cs{cU2R"}5e)t \TpMAѻ4@b??IØGABH/QK(TL^-I$)%t:gPV9kDNO&o[wsZ_?Ds_Sk/*WG]W #8 }=lݻ728m^eQ(qODcDEq>=1Z0刾RBh0";Qy|:zzKbDS|gKi;,Ut)elr558( wA;,5(DkWGz[r]y !x/LdsC[ %,!(c:9$>УQNK60Պ]URRqŬ6=mo Xm $pO?C׸sCf [8o[Ʒ[w7'=NJt}Mt6k.1m'Bd"n m= +f^jArTqWk>ֶk g63~DMnÌ/6;>nUH c RBXF%h,l.8#:G|E@e͠.ޖ:wk`GXnd4v3uREg !@OZF{/@ 8®bI&഍_D˭Jȓ0yzDr.q:z4Η}0;댎 jt^#ӛQ]rM(X*y 6Dpr*;d+QoZ+=TW4% x{Z:V;i$H<^&eKAԖAs.Hmaĩ5JI> ~J%BiIg]%GSfzy1*ɮQ?Vk&o?vrs5wMN>~_6S v_5o0F V[_ ( >MyY,e+ې5M]>4[p=}oL/^>8tjf#rm<`yqY[DmTa[w?4hA]l0ƌ$=<:?^smgBleϹRF}]ӳY22u{Re/8D;G:.8=0qu9,2-)X(Kq` ?bmO) oN3NqſV3S.aXZ**a7ۺ"V@[?}I)?+Js[o^ V!M]:4_Tz;VjemX.~R]茆}(CۨWeIh^i>$Ka+y>eT>o<ߎ6 j{u1pA8Z#&d&{_xdeƁb,Ko8|bH9k7 r52o%)C[$V+$gN伒6T<"(K̈́nrQm֭&fd_Ƣ. {b.qdtr i2 ǓT~. ڗ^n8i>'oȏ~LSQ@xrˀ%0p9s)O`~ibߣ]4J3ҋi*4-jg3%WupjOZN(.G$+зݧK.osGt$/"Fn[0vu>xj~ݺA3?,^@̮'ig| {<0ֻTIQGg Ecˡ^MƟ.Os_'~xjVɌÑP㕌Uí,4ްL?,HwǛFG{T{'{";\IwLG)5*9`JhYa$ctn(vDQxBRWcbY`r!!IYf-Eଵ1 d!ȼr,Wm8rVhKa|r$|fJy5{(͔ÞgO?s|Rw}vHMzЧxE+ x=QeY7b?$ "?_܊!^9w[nσ۱6{<ϤP_d?='j=V[fR!8rsf)9A>w*RFgEw~uI1Iy,5m |dK:D<$(k0ǕGHP5 EY5 hR"gZȕ(v@&^wa/a{ <)Z-Ҍ&D?&/6_bń-XU88YId͟ht}O)Uפ&@!U˙ĠLXEZfr/5l*mh_Z 6ubZ&W%SF_ۚJUq[*}]^ ./{CQ~@nx!5_Om"oy !6TTU a%BٞD %N8\ThNVci!*\)$Auqv gA2ҀR78QDq;tIB71G#K$Bgsbt8T@gsV^3(eJl\ j,6Ōbg $Fg)7#uK,i~#L MO 32=&l̓oTϲȘRΔ*ǖT5ۉg&&?葔Ya]Ɏ;]l;J`w8ԗ^wH_ՄX5W͹%{UN:Wڛ$L+T0A ՠ}n[%O QUcc@ ֓V%[S2g7\K=j夲r=/Wo=sw_lbSr[Oדv+7f` |<|*R{n2'MiԄFz쁪Ƃ{&$bXEт*Tu z;V^U51aNBP]!!BIJ) Yycu9"~-M*B_i=8 YfcV s2Q<5zn'xK*nD.bbj Vze5V2]tB5sU@/z9 B*2Р3c#ZeFYsN2pf)5&&+Ԝ4ε@@nV}ҫp:o~_PoPAhk;?Sio<9 |Oϟ}nGuͦk^qm4ѿV޵ϟێpe}C *GcZǵ}Jh;V;,tYPɧ)dkxk~޷Dd'L!zP@%PTmer)s,kVcSJUyM#k>;g49$-rdt !]4Л9ߚ &/Vw6d"}떪t{v7c"gyVeDN.Lئh2X PT$``"j\n:V ɻ`o?NjfIDWPT;_Xf 37/ 铬O鸏ǖUN$,]ӻt%fg(Jz.'hN' 6VK7fKƤ$kTI5F_sukd6YꦓK&;oPbSꀺ68+pg؛9{cH|Ie00˞];<).|~Ya'Mw~9]3A"2. )]%f}= $Tޥg{1x[<>'0@J@ 59Ip%ȤIߛ9gZVS7d<>5l+(YS+L6CdyKGV%Gz_F쒩Cc SU)!|q$^riѱ3N^=8¦i=zaـB! [0P[˩Z?Gg%n:t/6nۘ?>LGoghsP[gڡқknXZ;GJĢ$>/2$R0FWKe@ B_0 0D *Qk-U lpE" MAb$g )hX;7kǴGm"L>}Cznf!#v 8m,4I'W C#Cz); \jodO|ox1SȁOAZ,Y(UFW1 S(%ό!"B*({a`[[_B_8\ >/~n׻6-n1qOG6a@&~>g#_&?7lB~%]G+Ǔɘ]Mno`)<ū}_2BG}_/WF繪/< /fJlk PH!THW[,6#\\ Tru O\ؾ5{z;WdNzļ6t.&E)%pUdmĔ0'VXz^ToI~}x#mT['~xx+L!LDqa寧=‹ǪK^/Vc}Puz.g$`Kj$\gZ$Zvpb_mz5H;"U3WU/GW<6+̋p5p5DZOWMy/iv*=3t^zWreEy\}*"j;s ZYppլ$•FYEp%6WW\cx;\5+]y\GQ<~??|Izx;NVYUsM2?d䴾"m;}5嫉_ZVY n@wd^\^٫fI7UYyi\tJ &w=pZJDWieWCu[3\\5sWfYiiwW윇k t=\{5`]|0ج]Gi^\5_ \ܵUUdof%\ţ'jg.#q]Yo]6fU?~;+$ڧ!]M渷M-VAgc۷uאht´#?|}VI}<|Q7ޏW\l俕^ >uK! ߮f6fv6/xS3gSzvJc״<[-^;&<|Oj]}0֬t;t '+ߘeB*/Y@'2r5Yp>GAIY}OZĹ.kh*S%V:JuD 8b ij\m۷?z*hږG;[^]y m|@UJ4IJFj‘pdT '$ׁі>U}똂[;ƍ5Mj,ZԒ"4Dh2Q (C4V8LtaXrF>SѼL\t"$Qkusx P, x.cչI,TGW?^_EQI9mE7Yr&e5Dw <}X 7q_wu2W)KbTB)VX(;Kb fv9hP E߅^˅.EՒnD'SДAa,C)9i0 6#:ZN:G!v,p ԁ; %DH!MvYxEfRh" K&VȠ ^rz=Igϱk7n,L$ fJ$)!~ȃ\ GƢ#"kTAWwCIeI6wU&PBЌډl\j%?>C:b8k4Q8tDj@gSl R)BM~뭈ۨP&mw+$a:H؀$`F} ~M{-{D1dJ-6mM:\m3[tr6۴b^9$* B8 qt3(D6Q\ۘ5EvMs ,:T k̪k>HrDs%JV ]"z31i&0Y__ 2#BwCPD6HmJ F@bs&IU.j:RA ;(B ,kR9$TiOW,XoڪכaӳY@|?ouEXIBi'W Wh#GooP'. .F/)~BeFjJ ތLYz"ANi,WQti#ʅp^'МڙZgDn4HkҪU%(쥪Iu2 $jd-'k#kDvʷҔO>7SAj+v'QH6R FxXA¤VʣFZ&]Z -C 0!#]2'I{OQQ!,2T*JBU .: ƍuo1SUjԇn" sʨI8Y 5iTBGCtO]hEWrD$TiΛڻ׵hx;*:d5R GRy[\M3lH5 :y ̐VB}S}Rh1|. a7LzFVkñԿY ԏyO|>vLF\{Du7ȣ!`=xκ[xgIuibI &bI &bI &bI &bI &bI &bI &bI &bI &bI &bI &b@`- ׸!VGCZ+;5f@}1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $%HF ly @)8hPOd!hO峯jL1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 t$==IrON=`@3ȧ%fOC=J`@$УI A2 $ПE#zF$g$<9:3 @L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $3 >^Z>ޥNry_jz., w,Or~ ~%K k<ҁ.}p%GMPǒd=]u+t )I $+z6!0~" %N}u}q=c celE־U 䘗 W/~\7W Ƌ?yHKZHwvXl[]Nlf:6c(=>ʰ|o:^D9^]v-8RqP|C 8%H/Z?`|Ý_=ӵ/y Hr *^UqCO[X~Oa~}qx6g-S[^XtM>-חXe_RxO/6헯n?tzhlX<=-* 4B os5e-i&(XjbZotk9'6;E.}J٫^s.vpKu4Z#]^꫗9 /Ž䏵gou7q>J/˗SqYPS2;߉U X#?rD1r~OfiU*Oo?P/9]u.-K 6;K9蓾cS 1wjė4Isf!֦Njl,A&1ygmJ#+|GW=iMtLgg_@u.GF,+%ר"sPd(mҥdJ,ȣ8$)CݗDG~b#\}U1E+krLE!XԚLj3Ozby[okRn;Wg3w]Q7ZD RU)R")]cJqF4U}9VZ8Q1IX֋˷6V7nl>nܙc;0~xn}st |ާL+HS%}_j P¢j n,V|Rh&o}DD˜pF]Wn^JM" >9qMh\?l@Rt R NpU ZQ}˫:1]xlؐo?CjPqr kgi6@ȰȰVI};c{r^}*{u5]k;'`1NanpJ_w\]d⬞nwf M"j/yU~&XbZP%C2) MuTrՏLyģ5Y;b;WWaw|E??f=FitWn:/^ Ӟ6$w %gYΛEd)*1ݚ4m)i;ʨSzcz&>?;:򬬲|'g ]'\/QJ]餯ONn~}e}үO_m;qkqvqcw촾W|ԴE/÷ t;s~G^}{Φ5_#fI[Շ[QkWQoTۿ³{"sI2Y^nuͿ>>J۳6 iW@Z!8CC^ESC+/C} x'r n͸Y}Ͼ-w~9?]|^/;r5\cWgx\P\h@ ɴL+9W SJ ؓ EJ@}Cay#rQa;PԏRzяNEheu05^2wpgT<>A =$Q^caukDzi֪|zE x T%^d ɢvd|ʇ\;yb w*}v7;QFy(SCQ͆w{zvSz@WϽ)<$Z9%Ek*ұF kР,'/VEiu}(*;J{g9dn_]ny,z|>˿(cI1)c NP)eC"*QY R س:tBF Ǔ2v6+ǰQC ZMD wDզV݁g1ǹtۢ-@7A#*t` FY%9pR8gGtɁҋiĿ88F# "J#S=D&Ŭ1R)R d2K (Hn8h$un٬O{KoinRU _[d.yYw,OQ[<62I3З.է[dLe/x =F ~WFmZq-S9V+,\HdFYBI!Pw1ٖ:^4߆᷋Oۋ|صSAb-6kS#~Y Wsg[am y ]QoN{15*z驉{QoLVn}DDT:Ojy_*6~#sADrj5B xjD"^e|ы.9xf%Y Cx62h%U?F0R.i ArZCnekԚ>'Q H[X`DA"ƢBpqQgj+a|6:o,w\rݵWk'N}6IβYMS<[س@,/C~ri~aG7i3ݰRz=15| `b=9EP,~pA<>}x(.C1V=,~f CrCPwwr}J9:C j)+R$} 'l-윭 ;Wy.!]&SVr[ÝISK!'0JZ5袽& H&Tl 9c9$mtY&) 걖7EMAͿYK٫(c)m4"3`RpN`nv6ʢ]MZG(p3m)֙eti_A;Gɷ֏JMY&g]e=Ux͝mxN#UaZ6dq:MV(FД(NzE%d:Ռ]I }HUOѸT((y Krj/aJkpk(卲t$BX^TFT6-oG(MMl8ݾ_7crD JTPiS,&#\hr `J+@N)nF]?!(Ђ4tO lkIWٯxhΤXMkmv`wEGH7V0 Հ+2R6iƋ"$%TcddJ`uvkRѪ0ZR R1ID&Kw6%5&E#v&GjDjDF5.w"S7!X @1*HG_$,6S2:`nkV?*ItgxPi(MN m(e=>4};k' ;jNvXa# E0W0' IHֹ\ґV3[2Z$o=2!I)Jd JJ&x0 ZN9Y]&] g?\3q,_[%̈ym>O;AO\彏Đswq[c3bzfx4u~jһF aνFz4}F:" 49uM4!tmJ k@$Р.3y0VT,Rd>dL-*|RfO.9x+LDťq;0uˬ 5q!Zt;!_ci+8uXc37bN 'N^xYUO&P_j\<+^Le A 5)֔gN |L%dL2zxQuu /O`?sh1k6w]^?co{5npvu3o{^tO6{ʛ݄ݫv-n=]]wZ0w|wwa·AAmE{:N>΋>$+\Nd66;žIeznFA9O7 փs5+Pw!ӄ  N˞(ySHWK!-V1902s0F E:Yb:B5g4&(e^#J$( e!;P[ D L:Emb2(] ]j8>M{F)K;kūra[oMnYMy|my|{:"J 5%f_y98VWyH.C GE4ZЀ')F "_RT0{#cvķҳnէʂd)~^!Cs?-/t_~6zp~|*SCԪEy4qzb~7.ǟε cۣ?_GF㯣-woN,vϻ߬߼_$gɩkx5n{Ap}Hm w2\n|2slb@nPXl {|cPdRz3ۚ&e@Ul śjuRBL9[LABkFrS3tP)E(*80ZY/MrZǮp;~3o6?lq퍾4 {u_F-X)'Na5}mT}Ӽ^4_x\:I+Ad1bf€Ƙ j=K]tMkݒ{ ^˔>7moވ( b}3b#O^_D 'I-aJCaQ z}N#ȟkW VP "Bk &&9V TWNւ\Đ/:9T*ʃu=C8?v6c'džI~ |r_Mn,/6㨁?YC?@ܾdd4FC**q˛cF E&gQU(/I#ԡI0xD1no٘S Y’tN)z*Qt]=qk|B޻_}͜bW򍘥&!drevceGv/4Q߀ xRP"X|#R #?d<:bٯ 4SWeKpFY %"CH/{f/JX dyIxs{(6ǁބݾ)ln7t2ehzƦ6˟w ^T^laJ9gB44Bso X5 k׋˕ ɂߏ lUK+%:\U)pʢ<&v*tWUZ[+2J?|s4pU%s,p҂>xgJ+gj#*=vUuGbiC*^ \Ꮇ0/Wa ;eaR:wXp+p[/J\mઊҡIFҪO3S4>͸O33X4>͸O3ӌ4>͸O3=Kk#Ja>*.cIZqI#,'N9:"b5*WUï `)aRWL7Z% J8:.~y jަDαQz8(0V+q|j1 'R,FdןJg2xaOG#{Rb@62:'R" dA%Jw(bșO%g\Զ#BKDP>ËT)B"M(&{4e3dh*.9*}d{6Sż*}bA"E'^[,!!(@6dw&۔6bxcгz::_HA%%8D_fc( Lv $SdPH5SvS}U?[O遑c;nXQ!U )S.I N %FM?^sRJvVu7,mŗZx5I/E5O_9{b(a@;GLٜ*5IFň1:"$ @'`;Hos.AcdZcm*+NNSZ[9|:ٳLoY += H0X{u>m<7]*Ҳ*=Tr/,{v.X3&s5-hI8pvN[2FG2|w7]J2{o\3茑L>d*Ϣ,b!P:$'*@X$;Ol'\ӽ皍{ P[>U:RKd{i=#B6јb\1WT)A)^XΆI^}~"}o4_~:yϧfk/P@w洞< yc!R-]ӫ_Z _Uo H-lZxi1yB&?5HK zm4#Q P%c)6g7d)2xB"'\ˢ]W18SX$̫RV>PhC @z%MkG$73+='>sVJ϶/S텶~J˧|Mjuwsؑ@vo(1/1 $tGr*`FL+bZ3̑RKjݙc-%û҃TէSݎe=wmAMTiSPZHj{WhUc^vŻ(0>Eɗy|2wgONyٛaM~~owxCt-<=FC_JBYvZhۛmgZ/Ĵ3-tR0BֺPw )IDjv{a;>Їl漍pY"2Y(|0ڜS>ja3D&zC2aI&P0˄ 6@昌GCYBpJYIiケZ{7pYЃ !w|C zSiw_\y4WqSDny)._enlBuS\]h-0ޚT4AgCQLB"[ 1H56ǖ^Я[ܧZua nvwV;7ZcOQ!`1>5&ElVH*ռ%^lLCΒ[ ((3XcB9[bvLlfW|f%ܢXliࡸ!)N%*D:hR+Y"n-ME@NCJų{:^ѓ.I :NZTv}SfT+NGO~x;\z]/SE~3Đ5޹)YEeJ:,];p4J1oJ1g-SW{,Ď:oTaxC N좙M}e6lWy5M|'SJEk|-{8`Y2QD(γLo>a >,y/>=h!:[ ^Mݢ )U/Z=.ݘ2#5dГM^(jkW6eJ,}05p~5g 2Sd rQ!`pCQBlR|o+gJǐrGk tNf\٦Kmkw[3lmͤBX&0cq#6.m5 *k4_St㦵Q*DA + M¸Q(tD ?!tiݴ8'oDCjkf+m?d?KxNfwo 1R1mN}>]G(hda3g)+D퍩죧f߇Sdf:SWL?!4V8^\ls6X $])!EPH"uIҗm}:/+%Nʸ/mPKRSS1QCdd"I!gJ>B"YiSl<)N g4l:JkU^E34j"`غu+qv?~fL_~=fz؉%x K˯0h۷qeΞLo̢[]i6MeT n?)*PǤ ;V`B PjgRh J\ !c*!9s3d4dKJ%Q Q*Z*)T\PlyJos2h/% s?2ng)/SJ3 M½OU^Ý^՘놶󃳔+&_'q1cş0|2}MN] J4 1Zb()8VbK.rYM/C|RBeтd4 W ΔZ?h%.i͹<nEm]G`i!GH8_s ح-VIgZ8B.Q]%;$A`?l 2%%;W=|EFPfW5Ed HJߩ>,:C2#C lEG뚔E*+#(C*&FuvLȹ[FnU}шǎDĠx;pM7! @1*HG_$,k:`8ВuR'&6[tْVgs3d5bgܭ?˼ՁYW|qɮzQwxwőD,^ m= UQQ|QcDb멖M&"CEb(z1btÎÎ~AR7d<;3r -핝!mQqC8C*d=ÂvooqN fDT[IPM6jc$=2C@y[N+H;}f (T\H$*qC&Fx (I/=*C&R6SH0@h=xKF.]Cȹ"/rbv?-Q\+(vLiyɨ&4̓23 m]5z3ÀWj }1Vf $5=tٱ`P5xuanDER8t팾G혻|;":-3$# ֶ)j^Ǩ9 *؜Xe E2t[F߆ *9KDx@#?M)+CO9u팜-sgVڴO=3.:UL6#jL'+29V)FD$F6M5/`<1hc-U` :z!dN%9$O6`h6R.c錜 H8'a-|4(9XmbNڱq!ׂ&†o iNy6RGu1D Eɥ,6lit%8Tr#c=ujr&xV9\+IH1Hk6CsOǜ?;<=kG8u_Fy (SCQH6L&?wٴ[a5mE%EcV yIњtE5dxmd$R#ZQT\ڝ[r烧 'LJʌ{9o@1e12(RʆDT -I( Ri@zq1$/_uӋY-EsM.D^F~h(Y9%C P&:)KD, *%Z{A}q<$fa݊uDU;@*iJXcEKFy9OHF\–;J8?KF#椈(,&bL[IbM!(r!Zp H~!c`(gd|P㕄9f /n?e̖?w%}t9Ru0>~3uVY'Yߌf4quv]*6 v|2o&ߌ8tyq4ZO5KpUig%Z~7#?Ns>lY+Ta3`yAquy>Ax/ ,^=$٤GތZ/I}rX;߯h/*\sK9^Y}P|4p%O3||_U3;sGV.Ķ{Y}uQ>5;%cU7U=~wU/#Vͺa0\ ڳ$EY=BI!Pgر-u!4(` " )/z7DK):DH| L >GFufYJ&A"0H!;#`66LͺQm<ظls||y,ctCWl}$|  e=oh$2MJI7,5hո/x9˞ !AHoM `5%Skyb&2&d|=EY;'ɹWw,;|J_h;gcSwWp[IRZ_9>u= bytã篣ۧt5(pjȆG.o?m̗j|k#odr|˻ |mx~^Ws1'?c#:Է _C}~<:k`?I5rl'͑mвobj;l54+0v^~5OC^vQx)j _XM??`# iBھbF.|F}]b>J5R>g#5SVerdYkv\!gx}u47_}\Bz/d>w.^7G#/jYY_󽕝; mw\GOn#7Fɏw~n/`Q˻y{t<+#ѷ?;k (K`s0bn}\LWM'Y/OG ?^{wHV=RWLzoU%W}QWZ{uUtz 1bZ+&uDS (ucGިJtUY]U*q@WG]-H!W=oVW"ڧ]݋\^Ԣy"uu/*iꅺPWvPWn#uU F7ꪒKf_S R]]U*KTW tSolO_NJޘ VUsIhÛ'*BeɅ`JCgv!X/$ yB,LrF5ӓ8Ëݱyr(Ֆ$!~6֧j8Lv4m YFgݩyQvJ Ɍ%B͋qBP03ƨ SW+s ZxYgl)g++s k Q𐔋IFR)E' [GL*^`-mHDK\O5)QZRͣL Уv sOZ@*S4>k?@?v`'IMuA*P 1C )yP:{)9#ʘtYR rщUExT 6Ǵbh$e;b*,B5F$z-ub͖\CFoZ _f25zEv>ٱ-o!ljYA1"'2IQ?0uǛ&Plcj*cP/ z1;l.CvȪ 3ֿ HRoKL>ؤTU'>5mh)L~_aQdt0)eE#8z"gi>\6u_Z4h@5v8xx0;iR6Q+ڥf4{rvƿ-"dR-[v=N߽~eJ4^6YBi j eT 5kR˾Y_jiھ:o~OCXv[ՊMuਈFkd4JP5pӲMT iyǮ*(iVKFjV hrƳ7xN=JE+7\7Z0}OET="Tk=]? ?8~=!`rjs9Vwy%ѧwq+ v8c#co[NwY}gr=e& a䍳M:[2@ Rf.E(:MNc~6WRܻ|˻4 &!_4.G<-X&|`4\g;*7m|o~O>7+7uqk|'WBR6=ַ|c9rձxxt3Wx 6Y7WwRDb5U60[h="xVһ֓}].F2  IPˢCv1%+^X&2AT@TwϞ1LOTk:( @*#ɠR:[M7:*M]҄`AT !cHEP34؉Q'Eh!YצX1 %Cl2cѠIUY$C6DmPY*% : E=M 9lHdI$t9[[i>%_gݹUfzXҝ̮؀(?{WHd?6 40.qvloɒ[,vQvVU$bfF Ƒh"Rq$`ȳhb1$&zGTAGBTCO^c->H&70%Lȋ.f5r^ jdʨ,CR|wj8'BB{6Agvl+b1.zb,Pf37it9MèZi;j#uZS(ƁiFWsN%1b1 ~)x׷ݼ=:{lV2D?g}1N9 qstsUy~q-۾;E |tFekO;p=t6t8߾S%EcVC"5Yb6dHާI2qiQ[r+磧efg;˃79o@e+cdH)/riA*Ht! =Y8]0zf:4ly_43 lVaGP98&=:)kD&Vĥjj R}\yA"O*c<Q1*aΙbͥvx6RvjMGQ`Q#!D*(,&bLu!$Jx(^) B6ᛵ" (jT*cX4<1&aG dHERVSPOLǟ;=qzXx>Eg IO/2/κ.j\"OJVgםz<YV7]{O#&I紳?/'4M'>ry}  Ujs7/9O_0? Svo/lVͬp{Iax.dOz%3)kcgqm^u!E3֨OlW^26u'\exc/|w|:aSw\ X,j/JdFY=BI!P71ٖ:bb;in'Z <胤(D(˵n ?Z^oP/y~>Q⛺*QoYUw~D/,Z~UZ:<0@==YW>SJx ǾX ]^VOz=Es醵j﷌cK7Uh'@y~˴7}Nw?0> Ǿ%\"ǠLNˌB1]QР1 B-#&{kr,lǡ5j(NMI rHC! ʤ׎$o^:LОwDiFyPuR5#pk` xp R-5|],+^t4&Ss h"9C[Mu$B(Nl- oץNm S  4{Um`hQYGbQ{RNa{_@l% #b,*_˝E5G͖s=B[a|1yI{/[d,s!]ApcE?U[}]WMz^.ދ.xZ%ޭR[Oح ̓{[k]-[I"]?ZO4.'HP]v>͂2 5l1l5l5l(0X H/Rp.Ȅ7hҤ&2OkaY$MQbȠ0h jK1%rsb_*ܖ?:JWt1["Hnyl$.#y[ #@iC-cT:%U }YM{nvoXkq8?'iWv]nݤae:8hҮ2}vOn?ܿ_擻L!ޕ'!ԦA'/dw8=9y׷}WtG^q]O˚gtܸolFv}270޺uyznֽz;!-!WO? V.42.hEB9-{F5 1a VB@4 jR - !kM߸tP)ET^VƁRT *b(f9b8 +G{OݦZ3:OYlCЇa_ ~kB}Ӽ]u'*nNJ#eE+YLCLBaTݠ"jzT&ʻFo?x}5{vG1C\Pz;,)^ֈG#&E+k"Egܱ]ީb=G3ȏ1C'h LȨc= FGE_ٶu1 -1LE'507[E2Q 1c<3F~(vL$`}v3^ח_3M8Swa@L./Wc7\hTT햕cFHC؀EU,^>ȸK9:4)y{bL, ɗe*,sJ+JUh rk8&_vݽ/BiN +FfN o}`V5`ZH]ԅ:P;R 9B#Iy$ G8zi?u|6;b M4#{03ʖ AJD*_& SX䁾d?~S6\g Ǜf~}gu'~/}vavRJI\$dÉEp5;,W gdr63$C.DirJю8G^*(p,%jEJ+b 鲚tvsȼI"Yǯ3$^MMxgA\x3 G)'v+ -bK~49K3soFGda۔zVeG7iUh0 ?G?ip=Ikȏ"<0)A)E-"%wڍ&ֳz^6$&6?S} ~\m : m.6mn,x/K}-S1.Pmhs]>6~_^E{`vo0M}3^ z>kv*fcYJRcz8{F񭨙r?7"{AdJbUWʗbRљ7hR$tŘ+WRUS7WYJ.;sj̕s5B a5WVgBW2WJwCO撿 s _RRUR\Es4Ys Fq6 3g pMK݄im ou3pCQ+%DBΦAz&ndZ͆rsXRĞ÷7¾,p;gcHQMcǕ+7A]6.?}~ {i(W?"PIy9M\w!Re1DVRdvN;}뾯?F?Yy9 T&%6N%/%' 4yH >HHtR ;Ǎ4ᣱ&A omTDɨ OGtBN*;c̪k=q= /TMs~Zu*7Ÿ{Ӯ>ص8v3ž+) p+40~!Ts_ gbL;R a!"{[\1vDcǖCYT-\Ǯ{mTDĴE^ǵu{l:{K\zQc^7vy+B]m|ɵn|lfl><}nʲ'v#3}uOqӳ} FaT`0.Lvu;6qہ%u3!-*|՗$)8e*W3W)Jƿ؁nyU97y+SɁARt2"DelkuWx =+贂oTqy4ykWA]5~ݣ |C **)$qF̿@H)$។Տ+-z{%[۳<6P{v37Qsxph5}6hp5 x8QXjBeIKk.tN#ȟjcLhU>Y Ϋd,O40$! έ3*@ (n)…+`G7A0DgI@.R\oJU p&qlaXL=cޅ ߾L\vk|ucӤ)JtöD1$"Nz$6q @xE8EsƧdTgފa4U T)PPIt^J Hɂ",hM]FG<=]VleDpXÌKpM0D'@E h{c2HAo:sW3=MOgy^/D$"RY@L8v 9CE% u1"G.p݅A+q=q̖k g@yO2I7 Qk)bizuZl}G .J{v}[9IBK%xV.B͸#2t #Jpk B6z: B!Xc1DHKA:3^ʌ>bึ\Sp{Ǯ8}jî10j6O'G)?3xudr1hrG~jYyNMmg&m9oLA;'ϝ`pd!J_>_.6}M\syiɒe(#.ą IL&H;X'NN\ӾsB{(ހ9~FF؟~x,g3A f_ԗ;x8i A{ ʙ5VJuZz8{F񭨙AxntC/a=aUl8 gᑪ_g"mX+Y[%&hV_$1WhyE%̅hŝ}iɍVϙlX^lC)E=M$?f&OwG9 ~XƼbƪ٨8OksvƻYpRCS.u!)ԡ|'Fv/?Gʲ.4f>nNmpOQ@X4d р0ʨi)*Ѧk"I@Ǣh<^eBI%T`=o iRyR]PS?4{s1qvlCO&֚Xח~B7EdA5<3`7{WՌf㰩!k]r}GyQhgV1xIu$[n'G,hpgjnыqs Bz \)AG{fP5 C/˲ 8׌_"hU &mP0Y^Xϊz+Py8@Q*H;b TKp F3 0hi$ T/ez;_e]{oG*s@FGuwngbuIW=CRL55lɜLWUՊ(WlD5c0;vT]CTX)BtJ۝4^r֢HRuAXlD&oT/6Y3,b6I]ĽGmߍѢ9I|O*Tm[̶^yf I86R{*p(򭂵K-}P, UN(#M=o~>NB:M>bZb-lI`)߃Rrv$ >܈4؋vûw@~zO9AxNOp# Rx߮,Ut.g\9E'hD m@NޗzYJx6;%|]y.qD<0-qdzɀ߂6j:Z幉I8~DuDu7 ru"~ S]qWqݮtTO4CޟȕLA!(Oo+dmf)9\k#4rDˋhrQج$OXK<8i,2%Sd;RI_wuW  e w)q!X`Łhx*if.-P{u9C*U_Z8ax2GCD&1ł2p,YH{TSq )h:+.gc(1bD"2 T8..aT!)66LDyWM%&FfJ]B)SYGQaT}(KKW«m^iBKe`Rf-ߐ唳H\Ie3NaD1FD0bVځ_WV-m7ƚS?µ1C7M̘{Ĝ! I̹ӒnN3E GMAHRDnL=tdnRH\6 ɷg>^DZiTZJa,Dm*M )PNjtF/'$b,"p-˧J،ŽaP=lf,)F'av`nZPܐF$G(p)`@߿c T}쳠t- P$"\H\$g 1΢5:pof:8~\"AxY 8'1B'r1bTSdvL1J-"ЇRNpRp<12Pꁉ?swԵɰyREx>n76K 7FQ_5ni/, E|@M/79嫟p>uol{<>uٿ\ٷM%)mA>O7/_??>%xKsO,9lf͵RNDS^e I,^I!v-%S~!ip,- "I#?0:*yhqM9,(G(K θd nH`$O Ml:%drX.12L[xA: +y=.^P'y\p3饺N+?,Js:oԝ ]B.tI+ZU{<Z#I}c?]'rh廒Ip?#zf-iE՞_ Ǵ}ڢG͏r|;y-M! &^3w±y1LvN"-ȣ`,7ܓUepwT}fY8V\_Bjm!U(N3-Aqr^Q<"WHfBcNyPS2EoLQ8qХp▘`,xr T6vg{!dU(F2۹|-ׁyc#clS}.7zTӣl; Leg2TZmEm)K#<lj,Y#\(;HYV¶6قEĐ.7LfQU#~pt oGIX_]~6ϝ˾ZMk۝} cS6n~f>̦]>#O$yiwJڬ^omھjDhuڥnuZuA8yoߧgl;tGGto]v*e7J LC= Ics1;oHHWrx5^W?^ߣ8C[} ?Z{vs5yN fRa9wR {ux( Np%/*>sHsIհs"=Yy;Ҳ*@X4Fp&K;{ >UR6hnT*KI,GnAy༷VhM2GwV:.:W^F4ZZ)j;XoQgevlY,E=XQبO K> 7 &`ƊR5i@!~W$"pU$0TVj<`y8;/xt꺁_%TzGD,۠O)'~ѕo&z}wL.n}tpaW*l/v^*?`T)R=j2"IѱIK40IAˆ;a=2rQuj)1fYُN#, ӓT&fi^Ĝ5q(v"#U2V#~XTVI2vXn(*϶3=}T]!T;mnbL:ƒ_ <}Ex" i0U=_zuc*J/W|} {)SpI76[ ;1x2{=|y:vw]MaipƗisEj6xsU䣹r>x47J7}'1 Κ|+wvd+mH__3@Guwcwk`&GLiH*gՏE=ZC(_իK`-˺b%w_~&C|ȣߴ,dMxup>7U |5;yK>X\6 Ԯo^BcWÜSa] b{D?` qg6/S0x0[/=p H()mLsS3)J9΄9)vRJI iTLwg.Ο8yuBݯpmI$ρ h;ZS];l9FȁK%Ӛj@.s @!2TUa;Ɇ\9_G)1&R ɵ>X ? ~0r6(q':t A}9B*<?+74rMqfTTHuNҪ*$?[LĻGWRLNPc{f0V:a~Ai39/&t)ݲ [.][2!P**'hmaMV"lz~4EKբ\.ڣ-!jZUΧc K.QWku,M VwlZȬgmaۺҢq/ǽ]kz봼2[lm&ylWj!8B]/Dl(,2]v aŗ:u iCY|Q3чy2o؄=N:n^->Y_~n|ax5x:^_]9Ep k9/Q dų(BpZb%6M-k,g1VEA%WS8 67doN?nvðBuY(dmh2.kH kf_e>[|M\>UOb uLIf'9&ŝ(o(:Z[Ef(NWljK$ݚ(@̹ N/h$AΜMP)[ƅ#6F A&lBʐ"FL: . ׵*C7 6ݧV Nh+8mz.SS9&ف &$|cH K:S@elJi.(ށR'ZW%(XP̡~` H]x}*ĥؤTADVYE, ]}8T0² a?\ܛv]{fa-4YN۬jq%Tk5uY 5L P#V#(X'.x8@VsnOAmsgg !&0/N['7(Ћ+ԙ(?:ʦvF ٖ!Ynz6w}O|7T5eToȁc$*F-H$&ko۷P=’+>?DZT5rR7YaK1"%d(Y ?eU ;RƃU0ic2'D~.]NrB34ɒZz6A`Gȡ %"˄HGǥI@o 8N-y[X7b`9beSc@>G[ )S`|}Tˁw@Hl2ҌHdk,V*A%cL7]QA5AV ՂBQwTNz hKB\45sT۽;g0Ny]EX<{9;?=KOυף-u:TszGIVJ;=Tӳ2n`^ܲ[ -F&Q\y¢Gc];ߟ4F?B})?X$'}jz6J:(*[>9b=߈=Qg%G{ r ^6ikGţlNeV u6m>)Og3űo#zY)NBVy >\[>j7 ~}TLPyڿ^":o٦ EQF,/]$E'-dhEn ?uݣQ~c?]oJ#^~lk2r&^2{/D*v RpLƴg^cdd|0^]n¤)p^8}CL'7B )L%Xͨt!A֬R jZm=`O,107_kn TiFI+o8Hx'ui[iBb:dI,v(VA01*>O|!$WnwtI@Q:Jh,w*eLm?,R_a6_zXyb.*Y_~T+ќ1l7ݺlx9yߴnTAnU͉,],9˽6[5B-[ӋY^]m2Jpz ԥLL>tmpoOKAOmW2"Hg"Om'6XT 5kuXOCOmĆґ%qW*h_zը&Dդ8Aj4#aOyŌktpSs]>+1i y[އ#DiwM曹i'߂c'Ԉ]o엾]<+Oq?UK ;'չpDfQO#'i^K ˜ާQ\t e)Y4`/g]A 8c_-@|͠t%}DE%SoZr?擼U1UnXhq^Ϳ~هG,|&_>`Pˣyrfv9OD0շls'u:&OkvTF痓h$zy]/[/?IW˜9Y'[uvp6ZGqlZxwQ @qY郳ByCۻ:~_lIT,PTמJ-he[ jǎŵ>xܠ~l`U jb. M%ǔHC^U)XS$%ߚFΆoU{.{Jy|W1YzA[mle!N H4T|IX"J; :(2..zPAePLK^$RfWgۧdI0YD>"EaXP#go'}Ahq8E6`|nA%n+ĜiNN ް8rۙb*ũp0rևSfNN 7Ÿ/q0R#V#:hăF,T-Iwf anPJF1V 2O$*L7fVSAgJ:B&1!Mb&& DK$fP!{"Hϭ841ɮxgP>RlJ3.l DDSi]j8[S jQ 03t(3ק! 9*9wL~Z0zؾWy3< bZLx!+Wcw09}MNDs LDe-VdDe7+=-T6p~jPu;Ti,*AĀѕ~$q(Tgj'NQF#zBo7D̖Y }BuxY V#Af8Ӥ-<z .G 1 brnġ/? xT1F^ lgNsZx^(HBs/N"J&gNќ ڛߥo~ x$㇎ͯ 9f7tݜw].a:ȵCdts: ^0oHN1+l]?FwltfC)k~5e-!wnyw&LԃjV.`vr;TTՋ_ %lL0#Un`ĩ\Q[.JYŜ` j"QRӮ9* FJ?\E-a}%f6yv\ߴ0lJCs|7Ϥ7n߇mUF 8#'B?! {W5qi FyL^} ߺ׿;$֋ٔwuΙeR0]QlCVNU|ѺZ7d92Է[䧷y|Ut=u_"9qM2BB/s'~|$#/?5o%r]q*]tz>S{Dϖ~Ted)I".8"vBN΢ym5Q?}oTq}~e/nQr3sR#St<Mgb|gml;~h} ⃬ tY| >6M}ێ:G~w0 . :o[s En.\l~Aj|'CV.Tcg_vz8gS phՅK\VʐP !+18]hm6ЩrnU>Y Ϋd,O40$! έ3K'H2[mNoyk< _b| nVh>7vr?VbqN(iqkI$ "2CH a2꯾(Ke;Ru xLgM2Q(W2U tbKqW9PZ?}bMt.rbf^ziK afY.kM6![Z>J&Y2/|Օc{5BC/PXEb+}& Nvb~ q-{%[WTYW'ɦ[qjt?q,v=ޟ޾>ۆ:3vz^1Zl%B''Dd\21DHKRaw0t%R,kF2~ &0=>1 Քǖ?~z{vݺpd8Q(^p >{%~H3>KQШ-41j)cM"%KNA9'ŨW^,61pI^F\} IEԉ"):52@h ѣfb@r+\H21%X̶V\Vkeo!Ҏ#3nNQcJ裔I]E`Pz/jDr}Ds'vmN֎)|*" 58uVGK 0O2PjA1/!EbHbC~-C锌r7?n8pniEg] )z^P4X"J)'4I*SDSx@GM:AJHL `P`VpJ0eJĄrN*RA&FHaFknnѷ`Ϝ)aH>\mQw[[-;>G4_rY0㬱@($cV 79ZW=+EO+y:]HEF >'cp $s"$JbDmz2'un2_.dj #3K&$O PddjB)=qBz$!LxS8R(Sz+uZmG ЦA= kn?Bx 4zh ~b|(V >}"傞۪ bU {=xLykfxu!4l ц Őʖ?LYPCj+lXJ1ocȤ)eqO`pA_-_~Cd0*A=ĴdI2qHTjMB$A&t-N,'״__q<{ݽٵA&jV:ܽ'u!){q|/9e;e%B{&n4NEfD4o K8)+@/U" " 0 &d". b?MeĜ0AxgKP9)ZEaZl8[^DHӓ'rey$סW~лUUǯ͉(MI[98ATxj bB8XY{dYNzch9B< 1Em<[7ĥ31hpAg_* ņTw|YNYI%k!s6 e_s`+Xa\9%MONH]u-wƻ57a$H9VDGmPX:1Ԇ]@6z("mQd*JfSK˙%'tZZ4n Z-}&#s$g! $FM`JPe)K)ˮGCH_>C 7$oM5 4I< " N$c hJ5r8d@%sjsI 8I",g֭ȟÌR"X4)V:K9[Vz +m,WEO R$I0Y!/AHj;oJh3!d_*'(}AT"6hZқ4 s3c׃,_FzZ |.U?JBELoh kiBK$Qw|Z+Y/sS^خGGz*!詼mg0%quX:2- e#1bd3.8ag6M~:h'x?[F͒RO9X]`,y^ 5桹ux?_*Rgwؕ72Ehn\%U$)4EPؿ> y;¨WVx$z)XC /VddHnH*4-MYÒ{ )y1yTO_FLPmzL"\35eyLfsI ^D c!rOb,CT6|m1ٱ4~t}lDٔ8"謏Z$4^yiiG `&tېؐOU5qj _Ēu8N;3n!&_߱Ӌ{V?% 22TZmE@ImSF" "x7Mt\pAZ"JyBf:C@ [=7rd}5ZsZbsC]yMv 1_ zn-y~+AU2-zt n7Dގmt0GkіЫ3ʙ{~*f0p}sC͡i j<[Z[){8߷q \Nǂn~=hi o47v~2݉V9|{b~]F+K1| X, x2:JNJʦ47u6\JZe =rK>b&s#w;+U^/j֜7&E}={IGPB&XIeNZ",(lѧ%qfrW+$نYߊbr"A*na* jlt9i }bb#t*ɾHՉ"&Z5nW]҃l59kfY Rsalׇ-_)T4bFVW#5b7x&u裗,-nFb@0e5|"gL nHn6\Fj"5#餹 C &-HGt3Zsvezh="K):}y{ٺOH$N22#ECuPb@KI& HAXg{xx:g?VǾr:*lK.wGO9|%shD_x Fk%{?[Ix=%Lt2teUwT$٤hFo$~*n'eEuO.ZV򕌘i6Lsb\-)9K껋}f6Pv>\s://۝Uy9,o*W+o{O=0#8~ŝo6@qc mhLN+.ѳ lDpaLIρ-8ʹʞ rqdqHpp6Xxk'GmϷ-UwݫuȍM6}s o\uNϨrRJ/i8G-.lI{7f[!lC{j7 g5?^PJa2ootCwY;]}$ʷT6+;\o=z%Ei=g_m&+H 2 $'/k4g_ɪ؃ f(,Q9oA ]RT%˵W1%sH6oV dC`RfMn6#2  / Tkv3 m~ݭ95h;uwY,CnxlT|]/$%n\sŹgY U|HDP)UcD *!^x}%yoZ`EXG(y sxhKqDëlRD,ԐLMވ-:)]N#ȟkMVfj%e uAK˨t8Qk@ crYFc%7j@$2qL 5gGw!ۧ87y#>vg wF@]ueVQmT<(%!fh%b"%YR-_ˁX %EJp\eY,ZG@lTpl{j7FEzőPjg֜, 丸5.D{ho~St.yE,_Y"^߃{|7Zt!`djH+ KM5 5N4̙K)doG)/^ZO)8X:=18 %l=es0Jh9"AH/;p[!kNmNtJ[16ofr͇떑/\;ߗޟ-Gɳ AKk41l˴(\c|Y{VwPQַqNGhei͈I =*c֘x2Yzzq!d" y)VK@`X QԊq]oGWrmFGˀq%`r&[H8A)P#r(,s5=UտG~(ڇwcy2\~r#;?c8.rk"Vbu1b#3LQ,5>;=XAYסEζ] bQk=Hc{&;K]$B]9qK8eJF$pc4B6z*A@Gd 5u5a%loƻ2HWbM;Rƫwπ?5VmR`xoa01N`:>}1걅c&z!Z#wX'ڐrWy3Q-3+DiGUi")Yxs;|vnB`&Sr['nL]5[v6( ;R[&i"dgqୈ߳M-hj#$l4W1U^}J]T0&#|цR<]>A#W{^?n+ʛlz`f8Y}꙯|gy7zY ,7,P6-Z,Nyk9|:yC=sV n!V cx/E^.7(]v۾ 76v^U~]s+o7W jqW;p:7ŷ?|_73*n̽48/Pϣɒܩg B9ֽ+)e:%SJ-q j2p+A ;:r#/ͧ=30TNg^Ak #`^Ő [4ADL3/R+@ 29#Ѹ!0XC?vA*N ;wrG/ɞ-/pJyCjfʗ,"7^S^*. >.I|aҋjPZtJtL#Z:4qZ9U? %%cN5= Ci~C N4׳ k7&(+9PpG!mLhDRJ봒9Q^Cc-8`I\1 O( rx+0tDM>7GQid,rb X:cUz PEAͩOKmTdJXҚ;륗4萡&O@FzR> u ;2"JݣD \@o${$:âF._P3gOw4=w!؜=d$6R(Āptb8*3)m2~Io/0nU)/a2R]0ƌMj N[Ny@M҈8s`2;<An&+Gҕ|c[h訩R%!% %h\~"8Qlc?vt&!jX;ؗxz3]B8, b}uic ]$r" {8º6Y<` . [|TCrJ-@ Ly>"SDeJĔN8S&&O'mkn^UztFڥ-,N-;yK%(at\`RxFLQSh0J *B#lBzMxk(Wes},ncf*+boQ}3O'XX6.rezӣ?Oy㾆$M=dQVp-^Xآ9wyOzܭ)x_?s jzvJK|48W@CgzB)J•_R@G,tP&A"9(ChY (ךYss#a@ۻ0H *S'ĩzqK /QTQFO\e=sUkN\en B.YoLs5?"#Sw~tطI5 9o~E?M+`ֹRlgW?~bz:+مhs[n;}[+#ѝ?7w+ j\o͊Kk6{U?w&rvqv7jevcgJ"\Ctc.ũ3prTIKGhEW+V΋FY'WiW,ٔ V jPş)~!3BBUF8:ճN>/MNJg?݊+ۆujm3׳VjR>7]Ok U'2 Vw-u.!{b5Ø̕aHq2 ٥dN2@\e.Ջ1W7縖O[u~s8~fs(vœ\=[Lq\rr\<\\2W0 ɘ.b2\hZ2WȰdUfS1W[.\e.{sE+N~:]Vo:ӭ4%^>9#_VݻJϓ-A,5XVc$FfOA\spUCh\|oW&> DzWgm\Vsc+ٙt Qs: KQ)P&Z`B\K H%/\)z `kK8) _տUtzVg>aAwg MCӋ\l1Nʁ:ȲrI PW7AC;0'rYVɲpܝ9?$V@M;}y<&1@.才h1eh, O]7KBkS=伌\Rړ7NC}BY*GL1(Ԙψٖc TKs?Gy9g2q(2e/"MigJgEzx^)(pP5`$"9s|і O[iā툊ƈeh9A+uPݕ Dy &g F8iu /;ce~p6\}j(N>}3̘Ouz ц^Jx]ff,^ f ]IƳfcvޖɭMGC*ȗcvҙ==G6O5:%.qt5'WWznoLܾ] 6R.vvm&Ԭ\PEaaTB8WOÏ < ږ:,qX%3Oo^۔&TX2#TRA#3D0h!V^鄿12oQZ9x+o'[rt\Ϧ:_W<:J^Ck=oǿB$;@GQk`lҁ&+.%ZX]ɒh˹y$ǚGbż%cu2iLK/}]m>d@,.E"E`|madIk3Q˲ܲeO_zlufWϩ>F'%Qjo/9)~ΰmG ܄۫9xW_ld2FC.xeE Y$MEU,^L>ĸK9:vKSjoKi <ͤH2  9 ەJUkpgyZ5{g1b)i go,v:g9{)ܧ6Z#R #caF,=58KP 1x FQ7H@r*V򉾐yQx|S:%m+ w3 ze8g{ǵ޷:)/ 2tx=^ab2*1" Y D>059)AP[X*K1>%`{(ЪH[*L֒`lpEcdI Dvd*>{,$Dcs-v 9o/XR#tGenhsg<]5}F(tn '/Tx- IHQ1Mň.cj^|; eXox^nx&7W7=< )(JҔbA 0a& 5B\y(هwcy*\iTr#;w`8rT$YYHH%`WFQƒFG)},zwZXEYCt*zjRE]3a|zBO#ԅ ŝ6JkXZ*h"'識kl!&(=xtx@$C=# ,,wo<9QIhK LHf*,T$3{q!c2˲pt⨞XDx9O`xK8o9ۀɨL:6㞤  (D+"dmbJ7 顔ubrnBZ3d+YJ~LCt^9Ή-:d؝AjJ򏙩}&e"kO|L䥟˜lHM|";9)J!j CI%@E|䥻d=VHYͥlDƈ q|Xa 7A%Dm(m !dA,\ڡs>x>xzytfzZTa=‚K&)hTA&뙦!Y{Uxuf%;%e> OpTuZʥ,W9'0d !de\d J$yq^TEz;$0•Zg)#$o%@߳"L m_P3nG=OyZyO;jAJ^vb  PH7DeTL#%T聵*=W`Cmo-:yK@HIq֢Yg%3db.E`ltV|bj 6_h^)nV>Ҡn|+n~=!' ͋<}m/ͯ/ɷq sV/UW6Kv֢ܶr={g4 a|- W4Ke*OtTA0ݍ H5En3$D fGz-dy:yd A#JČ^x2Ȣ#9P[g+^TfotT{TBT!gu6S0Ճm!ڽ R]ߍErsɸuex+5(anb&Y2:׃W/*B\u3^ȡ> Ulԃ8.{ 'r]J%Iv4qj(6lz#.'=du[p 9c`t +kuM$ר СDt90w!-;Sw_G\f)Q-_UfYozotT\4lTW'b:S$(:QkQ8Z姨EmA6`Mռ TQ((FR`Mșb6&mPd JĀֆl&e.2%{ծ'{X[Sʶ8sTt&5IrEfݱXGVKE@QT<ˍ=j&"]H"٨SBR.١6bhÔ cFdrXA:sdq唔:_lPR\-YО&fDЎmXeTCL<և~?x'ˢH/_ ERI)$2D2cblÿN_5Ly?d#l-ӓS˜}7a^v{w̽cMs0V5ݜ?#wC˶/F|* 2YCHW!&huDgک~j/=.-r &-ш7g9obXDr"IQy%(%>xä,H륈>Rij't! _O Q(gI9%C p^(=IX#("MRNbrDž.i\gĆlWֽ_'T@J*kIe$7-6r>.x&Ya˓Zz?/N%IE$Ln"A{Ŭ1R)WJ@&2|ք+Rt1d,>0!HCҘ$$SZnwO]Լ=,|>Cg krsu0>ʩW%f^.F4td7=XtS֊nzC8b|Sv4/7)4U7/xӸ:_TCc/S6r?=&s%ЧNKB!cry.j(Рb-=&{mr,'C=~z TcqnJRXE15 IpV yh#JLE0}Oxg2A I7]'p>\m%Gnχ gM3\u4 =MuA&QgID}9ԲrY|!zQӋ1aJkm#W8 "Y`Yy dcc9#9߷c;YNIubSG֥U(FMm(ଏ\m@H)mV־H65ZĤsYvP=6#"jrGxyP[w(4?M-v>23MClon7r_5w$?j.iEp>D}Εjϐḫֹ];: ﱃkyb!U?$]E"lQzb?IFeaN2rQĨؗ3\bRNssw1Z@sduKڱ^ו]50̔AL%5*UPS)] lgbح;GQŵa -Ԩ 1~m"3DMh"yH3*Xj *D hq2E,FuQTv$آt6n9 -fih;cI[J\-{hzǙ@듟o9yF>\R;Ylƶ\V$nWT>袩>:`WưJP6RP5dbFKeCc2 BWd*QYRVF tjF,VHkoNTod֝؏tf43NB쌅~³b}ڳ~hxGe}N?Or{=t}}jq#v@Q&'ZL tc TS(D -|P\A&)*?9pqWIFl[FB'UOjmM{u爲+\S>H$$!օ b܄ż$,\~;#f{;cL|ƠOeX%Cr PPX,(7d4dqL,kbK,Yv/ln!@ !`r%YPuеdkK犢<{$\(%BLUd6s#JBTT]ցؤ{_!>tD1#מcq>䅏%_}U]}惋Ď^p}䚗 Ml?/2[P[_-H.V ?ָX\Xmgn?glYNaG}%g}~O!7w2Qk V7g#td bUF[I dAX`bR1U*VN7Mgmö ?_Hɫ״\_2_,E/yvq+5Q0% ?]\< O7vfn>vt}=/=Š|Dw=/ras"˟Gۂ&w;@V,ꗋDT_i Z8zp7܁lqu`TdlMV&+&5.hۄEgt)G9&vPƚS&P9,bHYj[X`F"\-SG@ѨBpxLbC55Nwn9'8bnv#ȅasl,y/M߽{9wT`[ %:j͹tεHzʴd"$w~pvGf +g.KY#[l{V'& CK 0#&kSΏ]ڱy۞4bg=/gM@0kkmŠwkcB-*.X) l{X"jĩZ тhSm7 '/ Pؙ?vΑE9AD??Gگry?[w^>a͙ztQhhuV<`/~C}r 09 T r+Ƌ%SA˔l6dTJoԠK,Օ7!]-DjƜ稵1 '+ Sut8$FUꝚ[w[V*x훵`vM ܭ+&eJ`Uweys7x\mUV10 =tdAeؚ0B2JQ̥IG:Ƭ#vK/_G?g84MաSl0YX Z# tr$EVX9N,y(tXccWyƪH WKELVo%_:3{QQy*;NdhaID:%T!P>"D6xL'`Q!몬WW)9B 9] XEA!8S25R&..LyB +Oz&_.j+b/momrf֞;z$W7 <SX*Js"a'd#@猁s-!`Ns3 zvBτ<ߝRr6dl@b#*\K'S:Qм Ry]OO㦍/+#ǷoXjV\XR J9.G2uX,IcvEg7'a=  ԳHQoewwPnlNvK=@/ â%aK-T.]ww#?i>Ӽ6oh^f϶\g$һu\[FeWHGeEY|O{vZ7KxM;h}}F}crgb5MpUIim M<{x ['zL;#A9fl-̘Rј`Dj!||Oh[wφ!hrU;ZGD6pu>*Csa^v\hkۥMoiڢxQ| j>CFϮ(VrN(oPkLkK.,.XpybQTL:R-|bzSF:9C 4Κ'U Ɗ+-5BT0 V֗ҊŠclUuΑv3홨-4lАƜL(K4DELdLE9_&F%K2( Mp4L&#p?6ʩP#&G&!J 3dh8:2 &ՍjI'M:x힪kWF. %掟b,@lIc+B vͣH d2 Xl>d: ,ik[<<g{Om%j~~|Q⯊upHGH"8:Z&rK%j"4+5$AS/eRS ƬѡNrIE9efЛr-Ev=MMu'|2`RL?I6Y)Y<,,34$6U+ؚ֣cM\v;Qȡ!*s׳r7:]5i&OK:@zgWr6d)$fQZe([\χm'}^3Eo@`& %|M>Zߛ}h;j`J.䵕{! S:v*IQS*DtYb"Ι+ l`ת42ǭ!10ן/`R(RާA97iϓɽeyy5>\=4 -mto'|^  š;tM )|jz*z:hӫQ"uKbc& R)zu -q I]|Wa"o%>e61x>WF͐^QG<71 Q}XnLěQ0սc>} 7V?Q*F;t97ssV(Ji sDczIEWfތ*rMm ^ ɛ\׃GKR5%ƺҊYk3t9v:߃pGKЎ\bZ \0+QZa ])*AGoX$o;NAw'_cWa^IQ\ ,M/PtۃJ("hxEԐ,:9(塀(u'Ac&rQ(lV'g,%8i,@)L(hBrNA*RB(hx*阄:OhY]_;^$8tÓ<YJ@1()3p1Dz%Gv<"8 eps4hii1s" bDd0 Ն\iwݎC,]bml֙N]r6%QdLZa^&v$+vY OLY,], t=X*[VJrY$2'L#H1+@ +VTw5p[|Zҵٷ{3 awgz+RV0箆xd3IEG?& rCh$)7ǽ sCcӳ7%=_0K2JK) Z%ɂ!ID%SeyU {\Z,҅TM-GwMﳌ7aq10D"1$bLʲ k˙^B Q=sI:!P',IF'av`6Q"ZPG$GHq)`@?]${.?.teD${ ̺ˀ (ADidSJ.$:\$g4΢5:pof:?' еKB#,pų b9Dr*[Va2BR 2Af?C1KP6ZG1,j*`@/ZJ9!Hu̕tR{B ޠ/>Usp<+x鏃Yf6<cם/?mv5nWIY3a_CZj64GM#eq|J4jXB //9 hO2kjZL0Giݬ=jĻQ\H.ގuM d໼:z>FE̮y>e6!YLS]ӭV2uGLex cKlT4B G( θd䊝R 'n6K~t{2IzyFY>[F͒O9X]`,y^ -nvk>.cE@-qaHZ+K~ZʵQۢ%Vg]B]FKZzQ2χ̦HTWFB%|!4XC Ȉ&=XTtW%t~&\)D;қ›gd2f j "ÙpKTᨅ7LvN".ȣ`,7)P!g!u{L txxUB>Q%δH>:VIk+txEmW^Z:BRE0mulȝJ^&DXljIljGt]u% Ѳ[ռYF.0 N(W4Q6oa~zۧc\Vo_q}Q1LRwh6PR۔Ȁ@ ue.k".E򄰭u@`{n1dὡo,*jymJ51>ZlΉ73̫n]wh/\.iE?juC>},IO;nga>m_EƍFp6%v&Z:䠁PjQ%ao Nr,AlŲeZ(^HQ&44{!d2v m`XC,}Qk8OKl?{*fWv58TjRRG;h DSS)01QSeT2ڣ\0+m\8'EȐiA&C&&H&w2s>fY :Wlx<-[Κc*+}(ՕzK]oR4eȻw]Kt>,)~ɔtHkr߫Z*zRTʆA(m[r/Z1Q_  Gv k? y"&1|DԈ}2h 1mmjڒ೽،:|I~Ȋ^Zb|` 9 *l7)lh<ub9ཱི*;{9{9{N8{koY,0f1Eˀ[FhR^ >{i+M#rZ6 ԡUc:s/j"ݝvP p]>Nҽl䃸xw^8 ʹPm*&%kea}}Eu`g^YeD^?T1g ZgJ&*@UԜAc- kdx{m%J5zЊ 4PU1@%tc.RnS&tIGfD1օnŠX%t@A5 mG6KLc4=|+"";p| Pԑa/Ԃ)eBe)͊K/A𲇗=(,)9H0ഊ4*{:s{& $$| <;Vi/jeu7;O+Sgc]Ρ͊ ~\vw]z}\0uԺKfzwt '%zߢmwE.|[n'GeuF 2;R!5ޕu$Bȳ2`f 2 M0SR1[xXE"ѓ X>U:w)(s,t}N?FILV:Y REcgWdcSM^2 /Q1'I8lca6"8͜G{F &QBHƩ:Q<` ?}+9kfʅY훓AgoLy9_zCW򟇬IhM^vrxmތx9X%,h"x&UW BX^Hc!opQػwGͨVZRN˞_((>%>f]Cί^O?Rk+nVʄdeW:y}W޶g{6j~}igQ?6xGH)a|'N>}0.HR]O{^N~Vy.~ٞ -=reW- T2^ȂL\DHE*ʃul}[Xǥ~\G$ W۶[>>s-mquʺ{n%}]-8#W KWBA"eNTP&kh_BNt)PZ䙓)]qgùCpCЈ9bًg0{觓idEd{6[{B_.ڲY]_2;4ڂ,gM @FKlʔT1ۖ]Wݤ` Bw5(m3O>/Y޳xAJY mWjfy7NqU=_fT%/WGNV{}tۚ}=.%?`?z_3YfLܻ?&|1Od6CeT|]_Z2MbJZVJQPL(S2q[Y,8*Z<َ|*CB%EYD7qwvLkVh:rɤ5(tN5r*6mn$(D;jM JRHX6?ߟyCFU)nnK;J6(K.%SjV5s 1v\sL',Ĝ^if |vibt& Alg &kk A!31ktM̋S^H&,j"6FUxt Yg=_ `udʇ ^څ$x(<$bTJQ:I6Z0U(% @V4V!# d_ r鿦 #\ldV$eHs"RYL#jh'v҆y&\z]g/CE3Đ'3IG% 91(yM"_]_-D3ث- Blw<EM+J"^#F²̣^KD%ћVї#cgu3ٗdc[ݥ3;jYA1"'2IQ?kI2II#Fے=6`Rf/wܡ~NEU_l/-t)4y]5J5Fu0ƣ>M<*$wxL{bov́Rpt}3>%}>޲=픞C^!Qq֢I'O2@ Rf.E`l@BZSޝ.ϧنnA!}`h=tw_:W/CT]N{+>wRqҋռȽkm[ܛow-ۓ>!U5|?ڮn]sPS~5u06}*7pC7T;f~zO [hPoxcwޗxD2xBB(QEm"KYaDu;t` ] SPT#z"He0 xԭ x5IUOnJ1, ZSmp(S=]l~z~yUYe U5\Lџ]BxDFFwssX>:( YS< DdI2[R.iʨX$,AKr5iuW mŠNw J?P|.oQB3Y(V=_ 퉀[ qGRqa3ѝUvlS#"ﳕiX Cц#/jZȑsZi)l2X4cFmΌZ"W=}6ԐFX%XA]= H&Tl 9c9rtY&ٔQ3GJ 2)e3S}(6I,[ƌl8;D{ϰ~pCmF'{2*1-crί];7|V&nSC*l) -aNJeE'4%'^Q@!c*$<Ѣ+)DhXKbKX xV{) N,ZKfdlU'}e!6vG^]Wd%xypƐxuؾ:q1{wq%vL.՜ J M,&#\hr d9VbN)XV.P=!(dvor \aN lۈViabEj7}654J>}!BX!s°fج-NMlIIb1:Ll!32VtٱIYDh8/|H$٨ηδm6ިy*7=DlF{JDm%"Q"!w"nC ::1q3b0)T"CH1YJ)XJDB)uq6%N :J3#+D^ m= UQQtQcDb멖L&"CEb(Žrrq8~QǾ=Aml)aquNE |u];w!n~|G"G?1?YzߢKBov9ՆCTd{:hw2ː:*61Wn Qc=XlbNڱq!r&r fa÷F՗ӫb`F}L3MTVKQYmH ,JpΩ 1F,z#kjyM/s9SZISк{= wl{=3:1O8kshY%@:tFak;Yn;].uՙ۳Z߅[RX4VkɪHW[C=Q KF"UQZ]5ʈK۹%7t>zZNzvve[n_}& 4HYH#C~"lHD0* Ғ V 9$_tzQ 몐#?4[٬Îr]q(CA 9"{f̆KY1 :gyZGĂ}Ϭ;ЁUNgEKFyVCr6|H!%^05cEDid1cJkRl5&Q5Rl7k^@Q##e~!c`(gd%4+ Q+is^`~ʘV] _3dONi]2iQYELtiuv}tUyZZvjVleJ=5w,*J7?M7l1Bfm^z#'|:1x{3YX.gNQe{IEޜwh?.U>HKrkxeA]m3%3Weu >'UdK^bz_Z}uVnZn[*[?)|=?{WȎ\mHS E ~ze_`{,j-S釙Ѩj6xYX{򯌮/ͩĶ ~{A#1ʲg\2q~/ m.ùӵ8Lz}&-E͒rO9X]`,yJrKe3zcwz[W"VmETךXb4f.K]iΆZyA/۬5^Rr4ײZeF&Х"Vvd$,~O8ZrN%旳MYviԟJocG <&39E/"F`YT1BB*%64V%t8l_٤*I ko XӸ (3?5K IyjG#?/K ܎,EP%4v4K2'4<2vrg%Vܺv.RAB.nnv[^ݽACt~Pͯ ]^#gB;W|eӡ{Rf[s%sFnNpmXun\lyHf}LcK͙0)YG澅%xbAվS,fQ9on.)cvK%OFFbfY ZSn Kb$KȰrBPRhDI*č RWpj~T~zJ>qvsZ`nͲ|@Du0m9?Ymszr;.B2re)xVCB䨔*1'AUBV1}%f=MKzK7CFgP2(v؏oG~NX[х4Gt cx}3:U`8Q#(4655SS c7"+DNJF3M{7#gk:ZI"ȒKy%,QK麅/ 1 t!e䲌J nʅHe A8[?Vəw!ǏOF3nfV뷍6EM9hͬomT<(%!fh%b"',IU2Fk9`+i~(;\[6Af qcPWUUjo[q'=θ}Xwvm~^12axՙtm|SSGZi_zḫApaD/X=FH8e̱KyL~18¦ %謉%B7pHRA*bȚSCyQxs;)V}L9ݲn}e4Ϯ|uUbz8~y[os,=@İ-Ӣ$Q3А")#.DM\N 79FpPtSSr zt c<3)2יFu) )L6^),DLd-f3hT јGo.+}_e'qۥZ?ߣmN>XR)\!xAFrA{41+3ȭ1A% 0:Jfv5q!d" y)VK@`X($Z1bP R^¤Ӹ0q\V!{'C&fa”#s3[!HOq,xǭrEg'X脌5MhcCG PоNfvW~w^ċw ]ƆOƭ)hDbcaɌdXʈэ܍1w-i-"7"mƴA`!!@Yr=5'%@JsJCI}Md-4{,rJ,MLQ*t}&}PAD"^CŋI{ ;w l-'RA#K8d4j4T\/f{ 5nv7pw3VSܺm## r ETbw*UWyt*Um3w?s@AS.p|=^<.ċŋ!YʓTL-A.__e+E.o.t4_8׬ =k湅 *@V4}"7/)[Z!;po8Hh:P.21l|[0:U8c+AæVBOi{Pç-o4vJՓ҂T̉=r{nOTEKGRtCg%-w%aI, 5RӉ101RċumDqA<;}'!Ǩ6ˤOSnga:i92eάՑ+"f J*c6 UG'j?_ hh ]9BȠDL2i6qəBdJqkZgm} rr^W40jVӍjna}}-ʩ OWrJ ؝ Q*0'D`S }Vs;q 5@h[AsJ"߀AC#FlLqr&X9ؑSԗI"Bb)9_ u͔"Rv^j-=rs%cD,Tn$x$+vV|`|٩CFP7̰_S&P*SȤIE\ S,8U-e"D%H`kjXcӡS2wxKn\/穻y]߉(=[xEC0(uy+WiJ` ]5((i$+u޹U0c :ŋ䕒,A 9[`NI }eXe7`-t^K75{j&m$;9БBQrSbR@y;|kI\hEY# 8 #r K4"CQҋoUկIx\e&ys:v"h/=yLmM/_j$>ާMizvGʌf@_ ^~ti/ߦߦ7ߧ=܁N1-:n{ ˖/FaT}2UX-xkmykvp=lGNHK#3\O \-EW6\i x%P LÕ $L7fWo1JAo2q0 4'ȇGoSfe1ikee~h n4E =79(Zl*pDit٧@ c =I>:S:؂  Ў( Z) MԨyhF w>71zprd.)?SN?ؗ~^Y靭gzN؃v#&(2IL mI"tAr^%*e^7oFcx Q;Ag1IO-cV#X΍W@~AɁڥug?*>dvW䕻4\eyK("FtYE׻$~ _sM!\/z/ *JР lbFB[`h9 -=X|DA/U+6ؚcMq*QMCQ:įz?/kp.sAԤi|vioJ $J2 a 6y`N/oy$-SմAreM_VLQ9*Ǜ!7mB @xp T+և?X>BзKT!:e!ep'mȦs  XƵ*EcqXٻ8rW1n[,H H|5|x<3zhVKIm 9b ^ RhUr_jҾ?Ŗ5h"3`TiH1A|O.aP@*u_ԩRXA(\|^i\l! V5 ӹ 1(6#JY e)yb!5-ЇVhJN؉b%mO7~WNTt+Ak(sbx߸}oaKGicۂ J 6Y~ޖw(+&ԡMC$6VQ$TsHǠAm@)]1lA0qށ]5,kRB%!TrLIb;ª b01)rudOϑ\X}I}PS=\QjہCb':hV@JŕT+XSAhQf8[CgnRI$*1p @T4i,SȀ*d*.ۚ7(< 7 vLeUTYFo܎KvݖEʣNbV&%O].vpt_Q (")o4 E#=fX.>]O'x={?`Lb()޵GyTٿoO2 5"mG !6Đ F_U-pok9GtV)%5>{_}lwAA;o.hR.)އ5ż$̿|$ZNľfs.+U,xuTINѩdB+Dz$w'YSad=k'g8'cсSkUM:ACOM jQcd[ڥl,i=ZAkU zqh)?clujxT|\zl-57tJ=U;.|T9ۜWZQeU ([tFjQhc1QL`cK&G.B r.T K6z@붜߫ _yV))7)6mًHl b6(V:ik}ͳ<mYcms}yXLZi}j3LVư堒7R\ukGP ̐cEt3ZA(0xPU(u;l< -w(<3^)0'] -0(00y4R,ۄ8㾉p3jpwYgK`F(P t:3Ux_8ƚ)4"8;X?]Af%""%WIjuCVTT֞v.J!YJQ=sȺ0ŧu+m_WCx0wU{qѓ={@F(P7 D:^}DV.ۊİϑzLCŜu1T]U˱VywN9~,#s{joA#P,͔Ks"-l(fg9 gj7lQ!Y*ғ]^ #(;*9 tQf$SEEL]#^JzYq0pdN ߝe8uy:Ye ܆M:v˫6S߷NW_/'/cA'G793~NⲜ|_ee'NMS6KNn(ݼMn;'27v6j6;|Qo3ጎVO狲 hߝ|Z{9W'_ZU庪bOѶ6|+Ϳ#E:Ds*uNoj}7:$n#Dts%^MՃ6]ZR;WVFệr.d- l=AQͥzsloz]?7+E]$jD1OC)V<=uvyzuwFŁ=A|{QvG&.~ܤ3poD{Z3 4W)B<277|m;Wϫ_I=jES!  Xqpjef;O`$ C_YzJj lLa,il8Zc5Lp5*dk`oSR 32F[hgnt'X[61趜Ŵ6==jyv$эq~> Ҹm?j_LdQ?p`bm^SD@@X~ӭQ`6vQ-ztWjF"!Pt"q{Ns܏og~J{Fܤ>x}Y_]:0HPPuh͖(P4&{T&uvPcrL0dM V&k3fLp)po?k:ʨ[ylRMpS鈺F2(Ep3~출#&W)AH߾~'oO/W矷e;Ȱm\v r3:|Iз ? B`a]݊RN*]D$[uUMJ@!VG@or1H1pdF^k+v%RqV;w[g}zܶ'yכYwbf,8ӝY^|^lZ]x l8Q*j-W#PS @ɗ'jۓ֟:}.nf׍s;ju6M#[Ac Uez@o iѶSc%7? oҔ@}֏y:ywA_~8xy^l}w}t?8L{vȈO_QK~$|GSP *e"-,&D:S})϶t2<e6={Mly;.I -@91aAhc$HTUKز `N)cy\ye8N_r|ZODETrQ1%Ѧw#+?%kߏp[%%;^R$!TS`4kzzU]*0g;Yb a5ԟa3FXסE%CC&u(*mݸ\$d!TBg=% \cAk(S.E2PX{CRcqV5~?ŌP0%Œ3c4@kCX*`cIyq1[ ndMV?p7: JkN¸ L!>{㢓(~α4ؑhTfvَ!yB^cJb/cعB.nY>ZhuB?k5D]jN y]vCc=hNE6`œ#Aԇ-Ge}8goY>qkTLqkzwDF6AtB< u3!aLG((w 59R)-FcC'V9]b=Cs^h MQ[.pNȰ{,q"Q#dE ?J(eˍڞ\W˛ 6+N$۞9\KB%K6m!)$UIe;eC]Rv$u9z-=GRd)Lq~\cʆxIhVP(_sx*t9tu2wWЀ&7)?~})t̋=G.y']0l:?@PA ECy॔yACd>wL30Vh-Q;]^$Mts )SJ:^x^9Z9qs.w0r%jW*! `t[nl'):[J=rwO%bxTJ1"qg"6yDJFõ(ibtڳ-|sE1Z8j.Dc|q%b.pL1)Dzkmː4rl$]M ?⊝^߬tbxJ-R Sq) 'xiL*luQI<EE.ݸ9Sw-v}%UNl` $^%e<>0\"rZ#α@.j2Ԏ+=HcP1X) mQ$yǘOp`.g-%@<.hMĒ\T' >iH"NtPH龜ҾM,΢ ` t\0'(ܦ뙴VzX+#C e'CG^M` m `/)0b9ĭBBn)xڕ]Xd_,zG}!tp㭌1ˡo.\.|Д3E;,M6_\r84:x0r'{w/#B\F뿆En v5wɧˤ'"vlwX!ł<ƾg .,F(xV8lcZ׊{}o{FniDZd "fiWMX:%.<| ڄe_9#x0ǐO/ˤVzsFVY)6sg7cm4t=}X=oТfv~߬_)|DӔF N/uAǧd`/x! oifG'ֺ|؇1\lA -IJt(`)A61TT9ƫCym^϶+lKX0G89N R%"9BOMWDm :v5$WeSmp_va&Ey$4J@lr0 T¾bvwӛ>ˏ܉6~xk=Qkܓ}f7]N=]x8}C*޺{37&sjmA<;>NuR'0 Sv'/hԎ.>,Qw5C]U %*3,(Sk?Y$g5!5}:qi>\]>0r:` mp3nʁa0?NGs; nMz;N.Fc [hg/Ľo6M""&x7 TaMrV}lj ]upu3Y]s~s0[n0/V?l6nԕbl_Tf Qz Z:5TpԠlB! ޻I Zkטּ(Lc:Ec% /00I0G7 ٠9i$Q00:Ss1p9FOpLfg$%UTP6% 뙱1$AY&q"Os):eY*Rl:Fњ~n#إ3aJQ_<;%M ;SQmױy1-[Oje?'5Rl('ROjI?'iiQK-(upGw%j}"@'Φuw`ER9o~n-8Gs_/X&X " DAÉQD;:`B >QH{Vv흸iܳ=(W\!oZw%5f +7HSѭװGtxܾ}:ݗP+rrZ3"C4 Y_-Y<5/RHɤ KS-c>pƌ %%0gRCs pݎT'FۗM(!-. L#'"`J5ppUF3WT!":7FFC4&`Me᭶&H̐-QRzłP%ma .6g9Ω'ڛ++uMѦղ|%R(Ku[l;rHjG~v$?g;襧Nٯ MmZ@֨`pMTh Rl/[nEΐ >IM'M@8t/c|vḿv[Ї l7&hvx  ky`a,)UJ0IP0[T`# )FJ1X5I4B厍I1ȪHŒ&"T*ቔQDVHrB95KL F'UC4hL\pUlt~N2#WQ.u?}aC5x=&1^N"QP0ƒ i9/>b".*+:Tӫzyw,:(P$]҆r9VDA LlvQk 0|W#O+k/'' [uL[*0g;Yb a5ԯB]XסE%CC&u(*mݸ\>Bx zJ.t޹n_W}^=tVa:N¸9!>{#8su;v$Yng@jx 7hfiRA  vX jyz4y:<{$?i#np?m?^j4Gi6. ~|7GW˛ 6|m&,qGmGŰ ,wFzH{,IGRA|:E6md[rBj+Amh0?_jY ֏˰"IRu{: c:B94~B_MvͯWV(Q<0&t~Hb6zkdP7kj ݿ0xNw,`Й!{ZT.ܑO`iEyBOhtֳ@: T$c0(y&]?lZׁ$杲nC?Ogznܬdh@.woj.峉}1rPi8m1Xo8&똔i"6eHTrr=Mk;IQMi./X9,9 7+` -q"j*1j mQauYYqor8NmTlW͔'/}>šqӥ f`3|DBH{gwTkPxE8%8ZS+b *{H8 woxanjK'—k)ݳVH:J"w_{%Zu]mO*8Uj: +D{"IA+UӋy<Wf(ËzS, bSF@iPZvkm^ol2s1 NU|vqiJFLy0TJ O(6L3#S3\ 3/3Pk7gAEj- VL*MB *eQ4YdK(m0ٻF 9ԏ*F8/~jSLRS=|X.%.em.g=]W55U(6.><uoFYdw!XR~ў-12Ǎ=*heTdJ(8FIZߊo~a,0QN&iQ;JRd%@ן90`!8QfT+'?'wxF*1C )ZEHQTƤϒXq|&1Td⨺TYe]xDfdj$e;b*,[m#_G:fK.!7XfÈfB8or,LDU>XFPv\?,٠uHCщ(ן5־/II#N,zX׷cюfrvڡ~AЛe5OdOy[>Nٕ}ҽ)( (ۄS %h44O6&YՄ8kǨd^y(SOQeDtFSJ=k*]Gk5E ̯A5Y2v2)9Q*J(j[[Ae|R)48qQA6eeȱ! ɑ,D֬tLPߡBw {~yg[o|T$ E$,sPD18AQ FeAZf V}B:+lI /"OE׶Y96%DW!0б=+BT p`A8}>ł{guF c= :0J}l\lّO'$g .[a˓?/ogd~)t"1UQYLĘ{KIbM!(r!ZPI*Q1?ǔh03B d2ZREP㕄9f _Y>SsH" |>c.^qo53-kUn_N.g5sd=fXk_2ݻZ QYoNyw0_hUf2M>Nnf͵{̫nDYlmn=imU.t6(c Yݷ{C ux bn|TJYQ}!x0Cw\ xPB;8e %B]Kζ/*Loϖ+WgyJ"{`x*Q|dqQ8{kM[F^P7hM[~^h8c/U47'mthFw0@=,d̪^]A߬ru+ZQK-7c7oN'>E7_lbfu@CCEz~JwO\0b%TiQ]t$lD<"j}!=Dɚ6[27?,;&8UmV[⭁?\*^~3A@rw?z{"ՋE:Uwp̅>)SW־MT) 4Ȓ*F[E;$?ΒPsu̗bQ+LD1&0&* mbIQI _NҺXlO煨u\>q6ٴX6`?(gޣ{S[~/2GرQyF?ڌQh;2m8rP]f٤e"i!:+iJ ]>gkB%HR! ~)+R$} 'lИ6΁G!{L]hQd];25vTUXA]U~ H&Tl 9c9$ҠN2>$*jHgҲ2f$Niٙ &1txor1x̟dz{fVbZw,Vg qOeru>ܾHH!|=?kqqo4žВIH4YT]tBSr82Aʃk]I jWb'Gh\*k^RKiwbjͦ42*dn%b!6B{g݌V7)@>8j}:gyr_tn:Y~K$bu'-HkK,r-% 8Pɑ2%ߖj6Tg'%tJ MI+ 2-1~j:O#vs,L:EmjtB}L-Xe=BB aQa #(ea("l@RxXu>Z=df؋:;ƚE*+c‡TL|<ͦ4Nm^,&)W` "6#D[DqB">&\dbHA 5BG nP & RDp "&K)+bSD$BXgZRNĤT)4V:1F8݌*ɼͦ4"{y(טbͤX\ԍqQp񄋏GY'bZhQ8=Dk#{O|2"(CρyH:ChG>y#1P)R]`#3}/> Ip~tNIHl.v6&rѭz kEʀr4zȘzo1Ɛ2:?mEa{8KW̆uDt_q.;{;p;3NoVCOE'>qսW>WtESwflrI$8-2kGߵ HTծ"phY͜o yjVUY &;jfT2?ݷfN2t^?f^xn?Q̫otG+]kjM{TѶzv񀿤|>?ZW<1ZFU-[yW+&Oˮ,W8Md'_:V<:pL`v ܺ3R3ϺכYۼWNoZDgyo43|wxnYqOL zZ^}ʺ_h]lޖiޮirh֣ %֊l{I c7`HŏX--g} ׺vYj4>n<LmKqnVn{ޝY4{z~+s[օT?[.w Ĩ+JBAԓ3GDM&|&;mrM>ivlJZrJQPL tSsUp?i]~^U"zuɧ+AK =Y}ѼIxsaI6k[82!6 %mQWeS=ӘA.vS ̬'TctcXp jZgrQA  Q[Iv/d>[:Z.iBR}w]<5 xA,_m f.ìUO_}xu+F:ӛ"ԯS_kڲby糟''}_F-;@Ľ5 NAgP-'h;oAw6!R( j}ov욏R GvʥZN9 ilՌsfUU̵`RXu*ƪH-tȥ}&d1*m@̪N72&QkV*Ċ5^ꜭi`žHfxחkFJ2 -}5y `CW+mϤ$2 x01Cc5Th*-Cٚ|=ltb`Ԗ17B==^gWpS:? ?d %FZ`#<%} m|2E׀ ?7MmV%nYUC Z9pxFYޱj @z`mu nJvFaPI;i80QT,Y!W&7Wb H4Heoڍ#e<kcD%mŐӉs4Bug3ΰ3ANۺnNq{ 3jYSw uܦ:LB@ƋԁqY=Xdpmg, J @Ku40n<&qΠ&acE&5-tD^4E"&4iYԄ0N** *6VYhZh{  E@YR:[Qx(nEe06 EWg'eSU_<9Qڽhpd$^tJU2 \I%d~x|46j紙4gC1I`U]|3 RLm)U$dn%$Cd.Xr0[`4&.? ujOW$2X BVI #jW[Y]ܬō-&&{N.E#K|5[wQpC,;Bׁ NXQ/Hٛ_vy6#OH;шzF#PUw=%gWJmH )*$%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ RUyPV@m)>A%)CJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%'l_W`+;=%P֚] Ԭ@HHR@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)>]%PO4G s{6J}J Ֆ SPM误'-a]<_qDgZbvo,gh\ڡny b._?KoJRt{s2.>.rս K *HHIYx[&vtT&:4 ¢5W%/ga#|vTg뉱i&W;#)Ʈ":%YժK gVd.y ?brnkQmVWJw<:6,-E{dv:tS·ƑyjNB){@d X$&0I,Lba X$&0I,Lba X$&0I,Lba X$&0I,Lba X$&0I,Lba X$&0I,鈅.ڊgGR6~y}^\ZP ;/giq}vxs66~EУڸ>=RƧO)2}#9Eˡ.κ+{Z08F"CD>ct^hk;U\aChѭ}|Bz8^ [3;ty^]IWogGmr` ݳ?>el~V/KBH^nwğ忏shlsvvb!R-~sG?$GԬw;2hy"oIW;9zqGCTCM8+8_mM:!G}sG}L*-/ѿ0xy%}ê-{dms?2o_N_K9y7*x[_gdzش5ю+Wgk?n͵jݏ\y6>Mf$cT+ǵ ~NmW9vFV޵%Bu=+E>dL_4:}<\پ?/FKWq1pq՝WM5A&ȸjWw¡.&B`R[1R !Rla{z.\| Hz~k :7*OVي߁-4=Y}gY^)lQ?򤗇U扨(XEQcvO|׍` \m'˛nqT@+#yGлoQ$)oV"P~ &5dz;Cl o+ߏMiJ5ǧ]ZIvȏ(գ0z̪ff7AxlixsL9luc Ǖɵs=dxOG4-%y*N^݅ew)':杒_ق;ũ\^Nu~icqrr1>vc;v|m߻qG쾍*WDo%qQ;>miﺋK8_H~`gnw &c=tCgt[8elF/|>t=vhr_S%GB S$M$L=$Σ&vg~h iƜbv9Cv(-LN?{WHw W+c?؃]`ؙ}=X)&eR7X$,RIRCwYɨ<""Ȍc[A%tF@-9>UFhl 2r:^ QHF(ڝ2/VKFtl!^ICqcB}̼$ LR"qI.DD3_IZewd5-e4{8:`^^ɦ$ \vɐ&k{vGKMO3 J{Q+e&GqyQ[ wX 7LlOed}-'o#JyјT׵f&1_^Qs%?5& ag#yq*Ԝ[sJt|0rE(*' 弖ue>OcGL ʟwix;v<\ao3Ϩøyϯ}M=SMWVBMmR'orm/[e3mFճVN0U9t2b#8 ܒf'0xe4KO:z4f~]kEҒ;w'E^XB昺UQCW]m?#[4_r,(opm?}Lp_r?^\ַm ?zqX6\?dݸϯOr6iפ|m?nJ9d5o/}|L}tE}w-u^i<ѱY_]Zn\E>-qZb&|Y2!HdLt4TSUb.uTfe9sUW)[Y(\^r /3${* dLL\ɀ5fc BCQxYfAxY;e#tԅdT;Uazm?̘rvS|ރXj7E*w6f=hflH7pӻ&Z6zwros;~ҙ3?!eiTTغsjz?yUxfM[Vزj?}aM[-Co|Mc)G3C0\UD3ʃqωppk0/@Oyыu` LJT1DkmBHK D5MsRRT`h( z1"|4X =hSFEL1V"70+֝n[!]c/;tEP<, ^[4"ngk*;3m:z#tt.t.t7ȨOum9W5X5z^=L ag'#~k[ͧ:jB/uAz9-M6BVh4ֿMh~B3KZa!Hu4Շ;B̮*9qk->Zme2}r'f 3ظnS7vG|ly4b,9ċV:Zh7׮z%rU!Em8}I z2DJUX5C10gF:ob*9"HjNFLTM<`bEԉe|ToyiR[;/(׍ќ}\BExJJ=I$`Fm~b[|"۠rt܌>ph7y=GZ>J@|;7ocHBNp{D#(쾄xTĠeE+ԚPY|Œ4ǹ 3wݗEOւ* (ys VySn/IfǦχ]ҝa mz XE^~g:}}= FI\[^ADyIm<ӛLn7Zw`<{%6/xo'^Y0㬱@($k1 Ph^|p5nYȢ1h >B[@$*,L$t f?^wZxuL+ MxsB-g/!\+xn2Wv@#[dNP e~"g'g}51iN)~ZW _?y@J׫\#Gfz={l~\ͷÊtuA]#dzy/K;'+ASI.⛼yQH]XA^d8"Mo~ty-4XbВ$%UK2j߹q!JȊP}3!imC(&.oSrSvH#tWndOb8ۛwK5M銒Ut{W80i9/)]Hb-WTkQs/6~~>9¼3w1=.Jm.[Y#]2ްܡᙻ[똠^{T\66W|&M]lZݷU$杲,-C?wprswT6=<Jy.ᗘ,)PhCrHX"E$I$pS }~aCy 6+p(:&Ox,у@9%h JʂA8KZҠ(́!Ts{4ݏ=.GݤQw;Y|׻:=1(/IP^iqry)FšyoHie4VؑQZ[֦ƻYpRCQV?Hz1ؿ = RrOؗBk־O>:OaACLmLL  N!4ZX@hYh-Jz0b @4tu0:b,6yak_F {q;: )ʭLfK^a3zM(o+B̄rZLY#Fy Z%9'I%=I $mQYͭ)e$Tx@*F%pF+=3(`XzY'oGqxBQ5aI0 &\ b;8 YZ~b<èdPDމt4XcڸX2{TP4dp@FK#D<~ou=/2"N#2r&3\H* xK*x;i%}GO&ڟ09>F]AĨ%I L (8L GqQW{x⤫=Yފދ a+BSF ^:m6ahZ8G4TDAƙV,aU%_̽z)9Uhaᣍ4t?M IJD79t.2\|#w^ŖۥHG19U:Xa`!]]m'}mkkɗirwEwrq],[tlѝ/{OzSN]-:W˸/እ[hp*J6.cA=TV őz T%.ݍS %Jp+1I =>[LtywCi GcBM܃ ?omT "$yTBTg˕ 48LF۸jV:@7TVJZPԲMS^M3ˤ`JSDU&km#9_enH/e\I]yH-IZ S=3((QТ4` gU]𴛶]D}-)֧"JJ?gAԝ5۾%P7VZ 67;/fuI!WR|^r\%Aԃ FXf(\)JAEo3>]¯ Jk'o6D:GMrM`.I"[Xwf> o=7sKZ~LTBA dDIN(㬍^8NF[V!p%ʧl> mzpCɚeDwI΃|>EbH;2,},pc*)#ջi F^T4tp6*>ktuzX^?vru17-nӥǏ/7x(VRWMk-F/*7tcfmb.V*+P}q(_}_ i/^ l[|1nM#{4@/mDeT0Ncu9*ArfEr$WsV妢~~}J<~h,,' sCeq2 LE8vi:ëv4uwm(zzcɻĬRMZC+ZAxh+i"MkDLI F ,wI -tod aCJU q}̝YK"Y4Cq.ȐnIER. upuݨ:룒HLWhiJs/!ri1܏ yʇ_Hm_FK-&!ivOX dp1+x{Z! 9GCQKW`1_{f*0P1 kExȶKȿHd2;T kUJFD@)MYhaH>NPh/8"@Xu),'mmF2f5Ƅ kK\3gp9U-70kt)qn[?|xv &~o57y"`߻Uvu:]Ei=)apZ貪e,C2D`~vSJ>ǂS{WJO,V,$m+]z| ֗hL(%]Fs,Hx jB-3`H If# ГhRCƔt 3gA퍜 MO\?mz㓭\ބJd.yYh6jNלj]Y-QY' ?E|-.|kMvm\1^mW}B ^öԷjFя?-4Qz9NB]y(M[w6 &mΌ "ga~Ըh_L㢝y:QgFzJ6Y70O`4?E yX!!*ޗw0jORTE~LyN] 18_<ݛW/lԒ j,VK1dN>f3~X>y MXep:q5fk/EO ^'UUމ(P^B_..zue;/J-x{^2}@M!]s dbS s4W 2$4g`C>mfQ9{ʾ! Y"p32y"gYņjI4ji1ؐ7Ec5o7f V_/giM6La|N4M5mK[Ѭl<9nkpn{S5ZGWF.[wNcl2z_wn&yW彛< ,h3̚EnvZ0~5[qI[9I5$4!=/!.:gG *eAmժ4w;Y!<.c2H.8+L_tpx` ń~OT/%?F"@3*vLϿ~}\nӒ%]Oz[_։ݪ4dG`X{m=h0EŖ\=Ҹ4(*Zeγgc=he >C_括 \a^ZAhQPGoٶ޻Q/}ZRLI s :l GͿ/G-4 ˜4_‰G ;'`RrGq~{qJ[jR8*5[ku:Yr 按fpE֗dE\6Gmы.>C{D1`FiWf{[B׿m'y1-/>kkmY){J#;{'Ӵ)-\e06`z\i\#YV:ksM޼RI+Ru/7_'<^|MܰƓ~U 2¥eh[ #I};N …ja7VcQRx] ^8;n~ooǵMttv4XqRv ]u]FInO)Y˖{4G")k*ӲFqH!l5ߵ92O͗S޳sBXɳX֑M?SMlɉ+#<[TĽPB#ǬYT.lӟKcu%ގ5om7#IïBjpy>. . ؾ`Qj&ؤFżQlR"li XR2GU~YڕZo]s4zglR,}ִZp`@Ԥ6P97Z :Fmt| I5K[[jVn›m0YZl&&MjBp- wOKwPE0cRNVPXt[\\T5[ϛbh!=B+;߄;Y0F.\"k_cF>{/.Ic `֑BvP=5P]ju;"4#H2](_ 26XAX3jVb{UPQAm|Z͡]K68U& gu%jɭZ A2&d]g(M DB˒M%IcAϫuJ+j+;.jP!ՠwVrEɸA+z4P@4rNcx&"fGܬFR ] !.)bD*pPgSO ŴrPVʤcB8jVZ hM m*3[!(Q.(ʨqƢ+x3`^UuA$E̾lD zFӈB~իf !1EDieR"Z8Z(46Aj"| C$C}(*мGw]+czh2&M yprJ/GF1/EuQ44u )<=B{h"/ڄ{aV9~/ u:E/xv ^ƾ mzZ$L>%‘FyUh+IW=$V*u2hyCs ~X|YE )|" LԴjPyUA}ڄLkptqX1GмL}$kIu < o 7tJrA+ПZ"hFymC7YK1Z1pߩicU0Kvnny-di"YrBUF%VX(;KSv9hdk}z /s#TϺێNFZ$n1J q[O κ$!H;V@@` )-f)c &˽e;VEσtD td..Ƣ(άIb&#Et4$?yP*ZE"8tDYUõ-ƢPaVB]$1dl)#ۉ6BkB M[tt֑p64IxtF+Ԁά$޲[6HVӥ[oU4E.7 i ߬n$BG l辠{ů%U SaCjh\6cs{_ymV0CfnoӮe\GIjd0u1vtrLC'Knc+i0`mi De Ú{)DzIhΓDCakBo1& )H&9%֋xyŴ aLw(Lj0R܌EEH,{U;:H9E =JW6"R 1tFͩm X{5= 3E5kՃ*M3| R5gҼLFALFzfj)+s蟐uN^Nu~kP׮BTt5zڍVk LW ƨsp$kN)#\H 2WB*-F`5>u"YdSΗ͜,>2lߝ ze=ޭÙח8ﯝ2 ӿѯlQA ?<`ѭ+mIɑL}աVQYԬAZBn\A1t:ݞh:!/N.қ%2Nvwm+k `i-t g'QeigÇ}ɠZr=&9Wv.H< bLʘ~]6a3?C٘^ 7w1.kĕ>e0o;Z144i`z]oQYi8&9VoeԀRJ X%܏Q\Q|~߭޷o/rFTiU›r -P K̢QiuFH#*T#k-٭" }wxWd1߯ T┆C]^0T:?rXןۿ0^@`?3pA;}Wo "V?-\$i8"ģ}ux>#+V`4^)fO[]2ri0ʴ1!+Z{ cN)/nV_߾n# k]?ыM&_sڇau.%*! 0Ы]Ŧj m[}`Z9e3"!~4rx0pъY^^>`@l8f^_s m׫NvXN4H C6{ܽo=RSFc.) 0!܇ZI5 zI= &w`G2/ϏI:V:|z>eꘄ^Ns'GNNj:ExEr>U v^mx+?] GG %Ljִ׮WkhYm.CE+ۄ_n1W CNAe@ʡWظ3&@ gח4~|~|aݙM]on^ZhwO"ˢ>:i2tbU ^|th݇"QQPEn6 "]Vִ)\=K%[#Q{L1H6]q7FE"gG`/ڢ0Yf]!16[j8\; D/K MgsP13lgc\OfnqT}B`dbN̷TYHRUs*8YC%)Y]y@6Xi8 &=JHX{UjVLHAX̦DGlW-ZbSNRk6UΎO>ioW~ܽ4҅RRթY}oxuE&C _(HEPkMFb,VoEW-T]RK*Gy8q̒8icvS8ktO䘛}M[sujoWNL7K>eZA*h~L>B -1ZϔB29$-uEz‡N`fmk3^m 0|CO|r|{5O@{PZQBFʒQ #t*Uaiװ/kW<$c1ǭ+\"Tׇ<`Q|0˱n'tu8ܹ!]k#IWĽ4,ܱq =ܞ>}#KKj,hʬxwI?M2?^)A`*)ep'mHs  XƵ֢9nŗې64"zh6-EalOm fOit1vD{BL-NX}7{׷m~l~}=2n[%n}ߡ]y8}rɊw۲D/̯\@zgzJ'>;"o1$Hc ,QBQϼGu@ukrjCrE4KAH.  ]EO4ZFqTgT꠬WC5{t5<:ϒ:c}jkt7Z(v̴p_QkTc|DT|_XPZqk'ج,w.h'Շ^L-S597ZQ/TV9oHKc4AGoXC7[w.=_?[̮pM0g0H_U0܏gj jQ$EP Y3рIC6҃.skQTIkQN-CԢpQج$OXjK{yp2L[e,E{"MhZxB PТfjִKVsN$Lp$XpcPSfb6e@9Kr#iUF뭊 Rm:/U 7_,]Rml֙v]r6O,d61>2#[Z(eJFTB<ͷ]R񡕦ɲ(T\&KoIˊ_ YN9ĕT8D$1"tk7EHϙTK@9so8Cqr39,缾c%_+3O1V4f=aydw=xV>;r.?wNXc5"-AZHCsENy%"rt蹋%XUxvS3OgMpnlia5f7E5tQtQNF̍w{|i[f^4kS" VQN&￘-%^PwOC dj44ZVbY1om[ɗiSTStCq0'|0nbaw~*ߵ^.Z@Y`y~Y;_GԈϷ?qF|xԪ{Is!Cyۼ Eik͓_dk9x4+Zk|;(E4sZI_ѫ/^\m34LMw\2.e!.9,=&8㒉{vycls.xv/ihkYr@) %+eQ$δH>:VHǒo+jý҆Lw!q!w*NL|C8?F8O*.bǍ}gÑ+rf#e;/ׁy}5{]X1HW2^; LegvhQPR۔Ȁ@ωJ=#˼O Ke%.+40+K b*3njǑX]E%L G/I+&X((P@ ́ـQfu[ičm֥\ MtȄ!`#yM瀢ȷ'8nf%_g5)9.vvGK*$!3Rd81DkZbO0U@‚>3STc!TGP\c'KGTէk)y@\xG S?JJ0UlknMB($&E pQ76N9G{PxxGwdwwwXY`&c,PdR9a s 9utvF(!W#W̽X+C,E-sm]k:8.5ɽDBZZZJ˻KCĻ 69OyK~p~-;^-uL{/ *Ц~98OAj)!3,C-T:p9ZBhy^sꏙ#eנQwDPԇ*+P_-s)pr[1K2'4<2˼&Dp.Tqsc-[֣ jd5ގ0mƙGhz^Rgp >zy&a¿?jt"Zpel@h۸,EcrUq%^𲇗B2%= FzVуfZeO{N9!dqHep6Xg~z$46g.|'sjn @6CQ?A_!_.f] fܝwp=s&~7ů2rϝpwc+|+ ]2텔7' \[ ќ/~ww~-yD1̭gȤչђ6_[ukn6B>?T`AsJr5mj(}t8&x\v( ;Cb)_1DoGLj+UZsZu9Yb:fADMlEU.@xBI J6,9\X Bg.Dr(t`tdMd?<(hy)t¼6y/<j.F-J()dܤ"Icǽ\8 MURb }))Z9A:\A'զ_k$h꺤x2X”DE{Xj:ZuEr[zz&j9wfC$W 8rcbTkP*cW[QonEufV)fA8BR%!g[*N鄼nĽsG&'~ ڨF:ZE*&M2ZKX8i/#+}J]|nbGD",")4"%O4ӏ_5JM2k/m/KJ᧛adb8g@,>_¨a|b~!B sZE v5Ҵ}q4>nǘ,(|WXh`FIC[g<(N dr}+<(G_OAܝv5ه Vh=j .mZũ ~tt)ȼ&xH*3 Z*hԻN98X*7YcBgq$pTk$L,XgyYswj!}b./xM1O7n+z,Ka%O}<#݅iFҋ2JF8Vg $ۑ_)َ|tFKO`%iE\P˦BKP*^vɌCRb7W1)r7{Áncx=[߂~\vu4Qux:;<\  -lKcZKI񉞾ER^|a2󶅬P&p5ǻmxK~٩ÕlGx;p6mg;wto@BpJu~Q_DZL-d|U:=a8{Yr7_+4͐{< _i6.Y~HCRܥnRVtrGނl;X{C{{5׃+jӟ< eX$% g~b QBպoHZ[wC߯bcli4lhsЪbf׳ռh0xLy,U6V_({mպtA <ޢ\ע3l1ocXɤY֡ҲߢGN.#: A+dxhARp :g%%K gE^E*&q!H I:޻щzzd8Ett%W xG8 fZi 6DHpj'"TP*) >,%jEJDe 颜zO׈X7rgF&ȫDwr+8',P/5Wgk5k"Y{=]/ 6ev'SaaEuy[CqP-ŇdT7> I'pWc;3Y΄})NV  B%4_~U6\ V}+- =#Pb6[UV{.$5 _"# ?[N~r zQaknW{Ri-RTj\yproH>ԻzQk>bow'p@5.dM/3? /slc9N MiQ;f7wg@lbo˷Nv˥҆%w&=I PVÁ*sZQF-KD:}nY_xNэy 6}w`w#WnxE-@jܙJ.J<3ˤ`RSD&ht_@XudTWGs)uV*6@^$_SQ~iE{4rlZZDQ\/x\UF .8Z=Qa"Z2dXnk"KG16˟sd> ~={כ7D?Q%G?.#i4gLtx7wg\ fTKP]vjl=ACyA]xxGuA"mpD(0X D$TAu!2 WJٶ46(I[ʐ%fE*x0 _ZޒQ"'Kkݪ;/_RP{N$yr$zi?e>>3Z^0zݬ}A26UΓ|BR}%6.iB0M:<:6d"rM""; nܝؗ"O2C2`@gS QYs,T9 DoTR lbN7w̷ ;P@ވtRg ]i2_5qM?vQj+LA$СC s&*wO3ԗ:DH9<4,JWfYJ&A"Re'?oTl{ =9U J}m狺{uen}3c1b EgNo?NҽwRJc^Nstչx9^^(R$hp:$֤VXS95_fB/!cBƗSŋZ-iȃ:,ey~6mYAj="z5 .C'vm^-ͺ[_aU'nO,6`ŭ;nxں[>,s2?"t\\/7C/{$y{>̦W2>}h}%tY0 7wR_JQ\xTOkGLZD`Q:tvS %E"m?x~v/p:9ximdVTͷ~߿忖n3 o`K},LvL6 #O_H/!_uHW_o)+)/"P>\]=6 wN^du'vep9v;C|w]&'9}7nyq `<[#Ձ&v@Zo`/w6\ow?"H1ow(,vVR(ǟ24{'mvro`1ʀZ@479SSZĕ1%p$S%&$%2VxUK\q&ͺs|#}lzY%]=Wn]n~@dFw? \}˼^zt_`kF V;01VŸT-~gm`Mw~۽׫Y,YȢ q}Dovہ߼MF'_핍ofw 'Y):Ôr^3waA$Z80! l(08elAEFhhMM0:Z-r&$op#S9HADوB΁:.DDj K $d]5DY=4^ƜH_{/ï[Ua27˶Wf2˟~2gRo 5-[jgCHUA,qw |m^M??_Udž#yqz=}?o[ `w=տtX5"rɬTnHGuJ(nU!}D'̉nʼ}tQ_{" |-KnYEAR JP M\[¾ wB@bu*[pI%"r>)Y{?F>VH:|F:>-~OYKThzI1lDRN~{e)uDJw=AgnuXm/uPt<ߞȆZfP {[rԳS϶&¸C*Hv;h/iW:-ΗMyl=JȞ/}``o܅j#9nROV'EkV ܥUu=+5|Ra&3ޕK5Y%ԯcggNەډb9vO<}NX РI!v,9η] Rt+ײד_Ւ%-K2e+vH"]U|]|*>rh; E6h@ 3LY*?5 &>lWy7:?$P@+eBTz !.kԥu%gsU0,d0Ĩ<}J̝ypzU^ϛ9^\>Uۍl緗W9z2'})at‚ a x% `ҼEz)6V6F؄˫QX3b7*8lfT 1CѽH͹mQb yN`3Ҳ>p}\:h`8 :Jc>)&_$ZECNߒ2"d]f24;C)妼9E1Pd_eK)]fiz$*W&Sg}I}ٛ6"},9yt *OR&]Y1oiEH-6MzA7}~l~^LgӞD2o~ r[O3]X1ևQU,~^<'B.z%Xq4P f$FM73a:5lC68Äh2dZE!{)xd9 *&D)p.sv5H>񪤱zs'kw 9,K.cHhF > ERŁvlwwnd7yw J@#vo^'Q\h]?-ٖ"dhuSw55q,Fd K +΃R?3^eE4܄`28y`1))J +Y*g0ҙrft>(f;e:c35SotUw9Vvgח]҅tMrK 1f'ԉERU:xtE]:u Du.9Aǹq&:=x@#skW<UWjyڂYofܓBtww}ChqeOLF[!|R2NdAwPH'-ˠ(oP؀aަ b~#iN9;{a-f,@zsp>kЎУw69ZފsZx+OܱetC|]ڰ5fUz:IS`BCcFkf) H=ŢQ|gD͘v5=BNO11xfRdn2a42kmAkX 0]r퐢Q%EczbE+PAJ7Cu3ށE1Z*AFGrAؔ.SڨQv|3zb!% %Ɛ Qء(Hi$JF3ST2CJy5YRI{ㆍ=os`92Y5c0'C`^fLCQ b9.\zVP7>✇笅j8b]"B3 ~n\+R)\N1 1z5(e&*IwԚs"b 1"P2ZCBQӈd"TlR+ך;ךS ]z?FT}JK^z&\~Q~Gp \,W|>A3vOE8XEyzv 9g-ݥlQ}]@dL7糳Wo6o7YnJ;Yp[ "݈=-JKtL6?j{bip a͒6kyu{qV:[i:A%5qwua* [qjV52~$St5F8젊9ےnBvNh?E=@EǤS$DI >tO$I={C#VKn-ygqvOA49pՇޡ+-=+AΨXz%lbWw'&wJlȮ^)vC'1-['PͿͯk}|H]f:yLL̜"/GB~bˈfytj6?!st[Ou <87aox}sLWm7nlƲ^_}pe8Ϟsxr*“UL@~G0_" /E~|B(6sl3` * !^o,!FqЌ⠱'9 9l-xLz|#6߉QOLJvX*=-TnAAd0S9-섏QES@%O:'$tQe@Uş XǘF Bp9!ӊ+ks{WГ+orP丶I8{@62iNOijf7WĻ颬=˥E/iqAeQ\ )rT[r&lȂ : @VmC [G_$K#K4Xn5=9?`Uł;;JGQ7Rĸ 26>EA)t9c| tB@hy Ry\s³ qRd,e"+ !h&M>b %NVjZ 2h6]L+ ^%acщ"@M4LK]{~FLTpK;oTfhf1sU/LY &EPNa/ >drkȡ+o@͖Z)c292G WAfi+0Sjru!&ͼ,OκQm:a~'VK{RZ QRQ C+%X-LzUM<|6%C=W|589A* 0FAs>X>LZ?{WHp`fEh\r@8 8ಃ_me=W7 俧ؒe[c ekf{5EUŧŧT4))-W.'.AO\߇v,쓘svQ4m¬xݗzvB#Ηl7;f4KCvL8_靖/(N報.8_|hBzeCbq!U,Ia4-#ܶ0!*MR[ >gR$'U 1x٤,P ['ΞښҩFӦ|`'=Xȶ@bYT&5\T,EFޱ(ڣVЁSFSIg@c4 Mi ͆ ?e9%e1Sx5Wmp[ Yʱ٤$*3c[i/FjmC):r ea?~]&͔F~JsڇIAETʼnW"Qr)2V"D`L-Z+h Q]𮯛y3H~cNg6|xyGxf̽Q`N{x:hef~1oCOE R j MB1 ${zmsw:~dJ1\YraLn DvZ e}Q֧\a{If?ڒ!d5%ee!(DAňH,ä؛22IA(Iv5+ چ덜.,'ߌkP^Z6,k!Nz.EYB&kr~ȁ*,zbߺRr ,6JI0yp,y}//+C޹R+FF0+C]ښۺ<eI1CIc V(Rډ(-Eɐ8M4~Ax)/$ƺƲ}Eer["Zb&b~@utd,kD(]5yI sqr1˗yRGĆ}N~^YId/Re(cEDp;l6”W?Gqh$bADPt"╓-%QaqRJ[ ߌ>zEV1 B5&pdLơEtZ{ ?Lk jR=CjwmYv=pr>-)я=x2~wy.||9AI&>,gS cr|7f *k/#ltx'Q+`1 ["Cø Udcpl8/M+t20- -,7ݾr!?7xfV5$EEp>o}8Y1.t>-W_xUeM~!ϸGu~A_L!sٱRQ~C|sAޚ=h,G(@:/j ݐ 5EP5 F dk`0DW@g3*c Xڜ}FI9R!8@ H$%x [Ӆ4Ξ&JOԸ];0uZ(bXUQ&$.rsH(@+>Cj{ .N=`0Ŭ )~J1c`%RiETЉmXQ grᗳ3,eߓNfrrPțQn^˝,f-{lZ7nOerqΟ^ƛ/SEΞMw{Є@i-`"e"P.YQ,x %ˤy6lI jyjCɦ'Gm*.HּD90hH͆elU2'5Cm!5fG{[Vd|/'?z} :}=߸Ŏ&'bu'X %-l0s $+hQ&G]m"U:{>IBJV..d4¾ W$ΔZF;L'鑋y*Vtj]c= ؝t&c=7&QJ EYgk9 7Ec 4#C@ʖmM"%E1dC*:Nu7͆=J;ꇓݴ-"`K-ȮJBEtX$Ib<(\d\J,@c}Q+2K1(ydVe=iVyv7 u2oA|Qn\l&%E.`; r^ e< dCd(cX!A:1ax x:{?P{{?G0a'6g_#b9m=AJ]%s$!]V<_J*GxΘ6w2AN$.'K.@2Kd4;';w'(O"\(5IB9jDA0hᕗxP݊+:)U 2k=P)r2ЭG]>}5H̏gH˻!4Rׅ kyMMx$Z0z]ǾקfY1[UgnWdBFaj"(; te?#H60@N 4cG=TU!b P]jD.=Yc&U!Bfk!Pj UgS='ZNgY|;œLs3L8,) @L#S?D^ePdgK^r' /+Z@VK5Sv3%b|P@'`<^w7ﳂ2B׫Uxۥ1< r5@v7]^b9{TS&nqknzw< ۼ\!wlW\IOz~hU}NyM;{=QU/t|~TƗG ܸkѼ5s{wh$|s,hq\7:MwrQzKz6H nvo'\5>gQj^@['v@Z}iYEV򮫕T~o5-z#?Zt0ZYyJv6/Ⴟn^6eOk:M?F+/_TQi'2}݇oX5W]Pw*;:D6GlS@Fƨ9ZҥØr6`vEU?׸X,Ĕ"R_!tZ%E^"ИSTx['17Qolm+^#V񐭰*j(kHh54.wl5sWy;Y4#31/%Z P>BNV*&TyAhQ2ʙJ-DT2E޵$׿ԏGb,$d ׸GL& QWS͗%Z#QTS1`Y {zzqNMUu'T1S(:;T**5IMp?6Ξ%1IoI*{z?fOD"[I)w[E/Е//1ѐ"v>E2BHhEUy!X|QLDbܥzjpBiJ1%"zYdRK$y 9 ەJUhݙt{q;ZePz2{5$?~?XeYd'ϾNolc0W0u-x$j2Bt (CP.G%R Iy$2|rFꉗ _/_O~ 4 2z XB#FQ7H=CH/6ʩX'b!xQx)>p~Lw 漇x7 /}?1gr<ڇ7e*1" Y D>059)AP4XX*k1>%`(ЪH[W֙$%KE Ȓ@`F_ܐ$Ecs,sˢ>ү FPfuG[>23:x7kx1HBRa_[WNHB Il(Fr1Ysɐ^ gvy) q!a4E4}IR,;̂Խek YrH)ocy*\iRbp`8_r}DAI"V n͸YH\%T&m;R:E(Eـu![I7{Nwa칕4~ &)ŝy,9!S*E8_O_/!-}|Ij^}Y4/prI_f_Ä/NK Пgygt 38M}kG3Ogp^CE=#pOX?ygX3v/t{MfglCʌ]bB˽Ϻf7Xl>_.ևRߕz #@q=P[}7O'zw\߸ĝRlXS4{z o guB1ݬJU.6)?Nzco}Y'9Na?g ~u_m>5Dőga >y ky;EnwnÑt# O|y*4ą'oNv!JdxP):NUJÊ|9`m-@s餜+3l;H%uJX,Q ǩNz;RAsi_MH=C VMP Q[6 asQT}jyVwY:?f񂭘zZo׆ .̒knjHAF 2Ydb\fl\/9vkTwJlܛOk9S-w}%9we+\OEG/02Z혛| ]X#ҋu˽x~0&B c;p.F 0VD"Mr 42Sgh)&-)F6bȶ$RHJch=iMGrydguZM6Г[kٿ ϼ'Nt̴y7Pws^|뫼4Էs.Q}|CoiSs%IEFStd YӋ )LsOWk04A:bd!!DƆ&Z3eCLtbTɔ]s JNiVx!YIzpN 'FB۔f* |DH5Ξzԫ5= tti` KY&T9'|Ju &+JdI<;g0@K^mӯ,@R #\ L(,eXt=kAa2a|B0@fP+]M4vz x,P DbΡ9;#"u $H̊`8JKm&-R= ?j]TI`p(UXB1PrZ*yFoZE@XfÈ>dr[ȡ'+\+DU!d!+ƈ)Q">C/{)zV>Ҩo'OȴIgo"/mw׭=ճtm+qOkګ~3 [C(rYf=9ަv"} Dx%Ps=xP8\LyiqO>wke*O|RZ ggFv:"w t{ (̀[M|6fv]y d A{/eG$dEGr^ThtT>nGk[P0-7zW=JHJ źySHz‡(jS}M+L'PPFasO)f%{ME^oƋu94J~Džb/MAԣ5Biڒ&N հn͝۽oCh\ۆ#-8Q  $dTZ IR ^ СDt9b >-'?fp@a^x]s 7B՗} 5z0[3AՉ 3EEEq\~^!+,FKld񵠶vSRQ"1hZEFvIac+ DR)58^8c|QkC62EdZ'ΞakMOJtνJљ$:t'wb!0J X*B. ,7:M^&t!xFrc`v )6L Ϟ'\ l:OU74())uXlaobbڀWE-4Y851c[im(V"T0e90!oNqw^0pŲ^YmNBb WP)ƁiD2cbl;&F:r?l'm5˓Z ȩUξڛΰl@xf̽R`NZ+T܂>ή[G.|=?Qt")u8.aPf1DBtnFOS2/r`Q:bIJ1JDU KT0ʐ 4bnۍ,m ;oUWif[':Yܬg6T!`TV l'LLfBOW7̷zxt9N6DkX9SvT銷dxÐ&Xc_{E3@8taZ+d )E12`> 樃,bUVO2@lZ.{˝5;')fPHN$)*TRN`9AI1(Eh|J+UZ}Do:hpmoFhږcw[8&+G0 ұFP`E MObrOU%]дΈ :~eKu3gHM*{" عh)a ]@L:etd-8?dZIE$ru"AbE)J)W< ߬5!E V1p C #[@C"8*1AIHZIKuojʘ=Cj<oϘX?rN`|qE ㌧ZC-hRbZ<&UiwXkWa/oW:?Nҷ/P4X]_sY^^0ͣt3_}jlu?% 9KO82%,l( ~-xBӖ#mwDI^uSGZZUc/t5};.%m1j~s|5ug:[l/>X}uUo8["\5bnKIs㣪S/ [q-Y0>(kdO()l..mYX,Kܜ-ؓ~9 vcSIeHBP $Fpzk~ZZ^h)y~>gZ{:_9~ 0Idv1_)xC,>D]ٔC$겺ԩ]q2zk*Ӝ:Ѩ7|`;hC-ر>זGŬf9hz~#|EcWm&=M3Kky~qdf?4рplc;LYƫ?=%5 z')L**NPL6YO=rpڑHzd_:YR#'kkjVDs1 F(m6bW6# vD\=-!|Tu7t萅oHvЫZ)JT hN]dgoOp `rP+f'SD$ٟ~{ʹO+k8䃞T)v{R5xa]txQ'RUN0HLr!Ak|ZtHʇE:a{_Lj$[Ds.Uxʁ`&GQhhM-6i),z=Wn9÷M/ܳIWZ ;%JDz!\'οu!ϣox#1ƿ&{9qDGuPM(5yHTtb?scu笏[s< 0VAt%ت]$ ժ`S&+ctQ ҃C[J(*pё/%G&s 1zb.$Jf 5(>.d0q\ GD}G?}82uۙI qUUC)FXˡ̂d:n+,0x@~~%DpXrT"\Qg6Ń?8O3%o=vPd4֜k5HzJri]lɆ D@jhdL8W䤁,P,tc9abJ6Mcoifyw4.? W_oSً$e? Zh.1H+oIU,ð\V9Ujh^R 6B4ق4يN8_%TH 4b(qv#vXP{08O}b[3q!5[OMAA% i0 RX (XàxX!E I#NX`M."YKıdb:ٍ}Rߎ靌rix4 e""a O-ۄ͸ʑB%Nce$غB%>l|шaO10i)jR  DCV2iVI4t1`q(ipcZJE50. O}u=\>(Lp*Ye"dK#{2VȬEeU.>.ϳPc*j<0^'xm^{{*}hqk.|I~Z1{]}o`w ̼:9կW!yi;b9x%뇾/[/VȭMtdw~vnb+k}y?? Ww_Px㦫kvwݪ kٝo/~bW;w^&WW[|w1c{핻㞋qR]Zy9?lat}= і~էg$ wuk'oy❟ |w5/9[pZ{6_f2Αpw?.{λFIzeܲgP|/D\P-йڝo_l5*XyKɠS&pI~I㐫UѰUtJq!N CI2',>3{@ YͲ98]!'Ứ;DF[Y=##mkcM`hf?#Fg#PNv;gnaGzYڎ][es)\fg?ANZR8Lq1@6(e?j0Ze=U'W|)Q3k)yU Lg|RgU{i?^>1rV O%qI-^ 8ҁxO0iƉQli1E2GͲs'b2gH'-EXY f/gzysuO}zπ_2/Gm*jocŰhUWWKJQ|zVL~whfR&_P?Ӥ.iRi@nf.)4__[qVo/~u>/~ݢ/=^hE+XB_Ń>5!ťw7_;\+|jIF7N WU8w4+m,|wH6qo (-&?~Z+vkg]֮o1I즳zA5VW쇶DI :vO !f6|㓝^4hu7mқ zgS߶p;eھmho0L=tx06PNfyD:eV특P?eH.$I+&C`#<%} +@dLQnX؆b77@ͫYQbe#sHT5DA(I2f[T坨䪧ֵ ":|_KnHMh>5")x^,.z> /G*Bh"jR(%6 *iX\(Lt~XXe1GSOm( AJt+Sx[ -!aQ,TG ?ON]iVPۂl2$ `~|:ſ׭ ?;8 ں!5"6_[V!`N h@6D߁K x1j$!ɷ%ྫྷbT5ʪ+@ @2IB=F n! UQXÎj8z5t< hwHX@zT̨ƌZJmJ#N`:RDڢ pd*"ΨN`&Ybm4 ?x2Z#vݑdDhÕ,"nT\1Bxг16#ӉTIXsMmtԑpiY xUmeVGaWǸ+,W039dE">&-J XҹdBaa-AsVnu9i9ʧUߦ]N泲e&AR,ԵW@7]px3 L[ Z}(xv`)jY:Z kH繨) 9! #v0ó=7fq\рH(R(."PPHyDB0J$5?N( j=jd}F M ӫ YW*ϕ+E3)ط.: ;bۡE 9RŢ4򨖌50 =TSXx .1#dnІEҭSq2"Vl,FjmR#5YH>9.J3)%&}*@$R;Op2ˇGH8? E[|-Yk8}C(1 gibe֏j`93#v"H-t1E H+hB?pȒ8{ H=B&`Pb.⌙TV̯^Bp4jUgC1IXU\|]v)&D,ܬlDD]R,9vP`- Q&,?@uj+Neg,+0gCm W[$\["\WD"pE+"\WD"pE+"\WD"pE+"\WD"pE+"\WD"pE+"\WD"pE%\uhuN+4< WR W$sG+"\WD"pE+"\WD"pE+"\WD"pE+"\WD"pE+"\WD"pE+"\WD"p5*3q_|UY eð-|~J"5:kkvw`d䓬\,aq׮yәElY~sϲ"\ ߵ[;~Bˍf}r(~\!R_Y3d|Y5KMq?&WWz75kUZWc&}QI4S.f:BTڻiTHr|Z6/m|ԭ&s:*UV.a=Yzy۪u]v*1~mqR_N_H}$!VޗGĸ'vUw缗b{a xWO9Sdcupk>[S0":%YժK ES?oI~Ie;lT!#adއP3`Eͬ0=P77МicVwBp^{8N1?e0.gXIb9r&gf4Ɏ"yfOiO_ٯ<= 0sژ يvIjo_6ζ&KݘYD7!֊brܭ={ $B'ur1tUȞ;Vnn@?-ln'v+1o5m//*/󟛎QN+c't=JvE\y hM =]~i]1C@}\?M^I|z.Cp~eR!rj^ :Ib bϾW14}6#V0vGICjAq׏zr2ܷ&n0.]0gq51 -46V^uS7ZڝQ,{m0RQ?mrDQi YcbhyJN;Ql fv68YoW ̬W-b=Ш:)EΉ[9C2r)^Sj"Zӯ *ͳR(>sUgd%zQ|ҡH |% 5JSʓ(Oj|yO^ƽ¼|]S3z;^n+<| L*EsPB}R9[ͪ(. a8ªBcDZ,7Wt}~,>LfasߗӉ)檷Z e]g߆Q0p5-7PȞGih*k;Je"M.H.3R4SKe:*e"nab|ȱG6>6mhHk\̞ _k`Ο!7dui~שXhruC|7XS^ᖮ[߲EWro;v0(̓KV9wk ;Ǿ(+? R") &^*G/6yNA߳+SL9ℱCjG(xr9½cN/:`Z^|\awo/gP~NG~vϣ- sN=^2.A1jJݰ},R"*t\+`4ulq6s5]y8&׷p^\yԲE t;u?PNZ8>|ӊ\l׻[ן%=nVJU۝#͡ ,wK?NOS}\p(%HJ;iƇTў axz {K᪗U8j%:vM)6-bH^3+׃ctT~D{i(JkZw 6^whwN_fm|ͧdǂߏٻ2΃*{al(t₵#<2%Cm*L>=ək>տ? e¼R #dӷRw[ىvEWrk>"ϦQC3]^(w^K]ĩ8R C&ZcJ*\};䍵h):CA5 2g1Nu4q@m+|24-ׂQPj凷8@}5)z-TCI+Q*\kϑ,댵$Q|jJsV٬驪Uu=!@OZ"A ^*)^-zh%bdEJ"H8=K*udt`P31)5tCQlqrcj^' J="ld.b \^F˽IX07Un+qz Zؤ}Xe8J 0˸rRanH{?AEnhZſm1P>hyIuy/ߚ]mq.>ӟS7F\Ķ >na\pKe1L\ܰ+C6f9/fsI?_5VfS.p9X1]mi K~"Z~D[+K~YƷQ;e?z,>zEZвOXlnʙU8[Z( e~]<ȱȈnzh7$B8&|Sv4/ "zgxqğʤ?(1 q ଉd&{^DHM4iUeVY"އ{Hֶ t7x8!Q$δ!Њ婔2ޕᡔn +Kn{rrW$GpX&A%>QF2a==kօ9B@2ʗ퍰V6Yha:/>ׅ5rWE ״BS;#Y̪vQGPR;Hˀ@m*Z2/>J1T \3d WůƧGM}fhFЏ*=gx9dXx~STBNj;*v;Q씇,<+FJL(N !*GQjQUAZea ;ːq"\}IT&BMHo^ݥ:t/T]0{;"2Cl$2K_&$8pIhǐqe 7 呧ip8Q΀& .Vk˫^D RIkɘ *na6Vj9,x5mQ3h{?MF÷;&Ŝr,)T/6\* ^F2$ѡK.Vfe(ƞOEmKS#7M&mǬ9bfڧsXc8=21qǩZUڮ=[Xx=By"Ms% N,3ThtNpì1X 8)BȊ 隄,jFMKp2!eX$aևQ_Nxؙr7DFk^#x&mHAQ+(IX)Z0[90m.Y*rƴQ.RY@G.zH@6F܌"զsDQkGH/ΚpuYKNՋ^^~>#mɑ{ɤ),Dkoօҩ% 2ӽ^<^O%8UBe}xTXDbWl=>?p{L)~D\^яv4-:QIB\jΠq3W˂6Bm:JٵPǓVL'ͥ~$ɖaU13me5<1B&4 6>JL*;c ݡƓ!8۝2k+w52ylMߔ P'C'ࣇ1Ui|Ϝ`s[Ԃ.$5>KјtW\ =eGeLAU V9ОS<&K2zy[43i`Yd1omif^Ǔ~ީ&u_~҅';iO=ρ nᅩn]9qljW݈z<}WvI{Bq0L[}2=G?%ϋyxuٍSq͚_ =zc>7w7 [om.Ѭm>?KEiKId D=Yꀍ/M !{Kԡ˖f]9; ׃/RZJÇ+Z?Lkٶߗpwt~ {&*{WRk0(*Zy}5.^з^5(/8J>XMXX|[ۏo> K5Kډø;/C+atx0ҜTta51-JsYq#&bzS,D3"`2愽F8F(d.)+oE1̬FE5jJ{MP%,(5%gsZ@:3KmvV81R(N).pLX;Ft+qH5}r+<!x}\oU"Z;haw{nj|~Ʒ*u+aKXd3ZA9 HH*C&eE 3&*⩤x&uUNkbJbIp Qb :YA QҊq@B5W橢y \qTR٦8.YsFNLR͘h1'C`TpB1& cY ;ǥ쥾w}֧b?jMEe7 ' 9~R %^x99 `wq›@d@p(JC X.?A#ޫS*T{NO*#+>\X-%(0n~Fh88j?]E)S< I&Q$mPPv"u} umQ~W&]wPVEz$.I;:_$7'.Ay dJxE_'aߠ Gp~4qsQp:d}V1ܴm]ăP/ad![*Q~wQ0%IWGwK<]=MPg{Nqiz>=ttܖ3nۥ6no/yJUȎ:*G爜51YFll*"bީfydIlS=7˕[yACAV* Jea!1sN\l@ƽ*}ܗu:1侾JffBȨ2aLY%xQ+ "g`a4JJp`W _nRoھ:Ҁ..켮rn4ύ0JsR)NHfKbs3cl  6fFt(R 96ik4ɞi0]˞>iΔ|\gGd@7Kyn=͔"0 r!) "iЊ>wR|`|Yz7gAzH׹+~v4wiZ4ۢև&o ARF5sWv;o"ZOxGdEe((jc}65=ro-h}o=z'@K}6\n}K0+˴mlќ CЩԃ n(4FOQATf0,g-Ąm eCtE/NNbdSųաZ9o:z@0[{fӯos.E_/gHy./ޭegX?m}U>sAS "x]*$ᒜE*'K'Nϻfkt1xSN)9|EFX\-AZW&f̹?oY_@ð69w|{M1<]W7,QY*7jasG/ j }\ "V>4aK|u_3p?ۣyŇm||Cn\©@7V/^m^/gۜ9l}JV+x2\怐ӓ~JCGJ~6 pb@@umQ#.36O-/Kۼp1b6*6|W ⹔1,]JfgqHZ ].h_tp.˻nHX 1߶E#'rx[|b JX|Ґ~`o4IJi ]k7tڨ mKs[fr6E_Vbi 볣j˄.7\9BUKO-ǷH+4sAÑ/>.:#tlEڋ7nףFkP3ցS K~IqaQ0$_64!b4+1зp]G,w=d6d(q=xJ$߽]ז !uCK1/Gmټ}^MeY pk$9\љ/7^lYU$-[erN.5~ k 1ȏ(s-8eT_F*&RQ&'(P. Y.ꎛ#+w>rL3gGrGZ@t|o2@W 1jGTCB1z_8Ҽb=A.& >sΪwtS=L,887[ Xk}YsV!ipHm>] g}Pô:ю`26&{H4V7l!#R!ok';ѺYsZ ˞xx|cDaHֳ`ƭbnxږO ppRd I-H ^>6.F)l(qs٢*I+JdT)j8K 3F[T흨䪗1(!u>+ݜx8ǾYQZ4VjH)FȚ,%C|]ԚER-H@"Ӥ'i5/&v$ N&!)Q)3mI Dnm^$Lȓ Pt=+ )D!; x{(%x (+J|.MV^O똌' ƒ-O`׬|x`ZS׶?5/Up;4p mgu&Jb _eMn]9 ڢQT 16b MS":+4ƅ%DCi|5lIՌ9&G..`TV!( HJA6̹(zPV \rlsEΡd0( dJAҔ^wvNck{ḈT9k US%BaI,# lʄDc$4vHn5Vj^@f0QfBnU{CJ `\ڃtQ|Pil[֣Y0393|%i*UCI$#s5Fex"f#fC7!v@l1 ٔ|Kp`gLU6U r VҀ'ӊA jo.LEwF)p0GRv<`M W4kϒo/Ez@_P[.()EVi+nXkx5=+)IE,);ֈ yEU@!ѿ)%RL8(Z,V +\9ޤ6%+:c"&LԙXA:ku+k7" f*y$' *JQ;( 9}2fئ3km67joY4]aNTZ$ >:bu`i؃ϛБJ=E J @KU405n<&q 'ic٢YE ʃv4I"&iyUB|\TT e:?m,YѼ8{ *:v`:GMۀ8ڂ2UɂN~d}%^ɩ⩭, ) T9%B]VI|'ӧU>}]=\v$d*"u1ZR "MG!߁]J6HmA/fsP6D6y *0BըyހR"de`w PJΚ%8&oC@EK Q DnA\ds۝2isg#kFW85 Na8J"mR<2GCmg5Ւᴱ6 3+H@R$&JI>2)eCƠ""*+Y;Eܬ4̹(c;"gF A16gMl9OtXfHgYtgi&ڡ/ ,§ЈPJ9SeQ{ѣN j orf .y͸vZ,A`sR`ⴱ㠍yyc7rIbaf" RLȎQXlҼPD`ЅKϝq9D LB؄rK\P{!((;cgD[%=i,^m!zz7^l`bDIKm("1jG߼ %KzI tKa$06 *qF;vb#Un{÷|+~zK=5i4Ys5+]Jn(-5] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@] ԕ@_-RLvo@\6jZܕ@J'kTIJJJJJJJJJJJJJJJJJJJJJJJX $J}R)HR{o@J=W] 5*zbJJJJJJJJJJJJJ7nI8 37#!y\`1>սn'+ݶ[mٖ'aDdUȯ@ @ @ @ @ @ @ @ @ @ " X2gɪ@*<&$׈q@ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ nV $;{^@o뽫suv@?m.i@5$tC-]Y23*ao>g3p\bDMi 2Qw͞y.f9Ab> 4|~S?80 ;+w ]ʛd`a]O;~zxcb'm|g CkJuYhYn2_t[3SBQ?)W ֘<OrZ +hn$px!Liǘ{(zY(i{LCN-R9H &=zl /4 Y~\Ug6˄ (T+C#RJCD(X}8CGg8`>ٟ~ k=VlugoW&LO:Lg5ӤTg47y=ųGzqbb͏6qŹ9:*("x`cG(<_uN=>AA?7o^nOh]OM!lG ӂP̙x GӤWRcOV 0d#0-ؖn|u$нvb[F ݈0pcCFL2T:_f,j¯{5ەmk"R ۈ ەU g 힮O9bקSKׂyUӺ2nbώ pA"(qs/Ծh vH!0u/ &鎬n'f'EBc cw XF6}h/UBxPAEOjjUY>孒Gl:ޖx%k:up}˕.׷n)23*o!Z"8zmm77ԝ9̡A Wo۷[LdVUؙ/7)9\pUMa<Ҟjmf /h 2|L~lrל%O"eU.($ɽ| eY*녽ՆV䤯f8Gv~|ɂQ1JN@pPy ]']חf㾿 sJw#LJ;_dV"VWZU ۭqC`q?+~,͂xu5?]h"l DL >J\s 0Cu<{yFcroq}*?Fn}0&Erhva}]Soj? c:fA׀lkVb}zRW>~ޢccqQ3_pĠu6[A峋2 f!ʟ{7d_W+]Mͺ}KBö E6-{CHK/}pH4eFP=ӠX)\IGMb=)}^u-ztv1ۘS4vd713؁G394 6In/܌3JCh" S Dy|9B߭! XdnumuF8JNevK7mN`NWtп_OaA~,x>{>}~Q3wz1:E·Po^N^C"l_c$II53C>Hxy! >HJó^^MHx{2^h<A`d%h-.;8嬮RndZB_sCKO$  ^I t {jA !|BFP,P+GPM37O:;͐qZ,o!< n)羚 ._e< ܼGf =#z NŠ"CmGf#ATY6n Q5b̈T_0ё쌛f #cd6R澙v井3LN/1d Ϭj㙶""#Lh'CA!g)3J"zb'ԗwbhmXI9@ #1kic{qBk#g`sjϥ6'K=A =8i*6z׈=^&yO#QK(G|Ц0sv#ᔸXwO«pZ|/n𾫋gU#f@=xY0S@p=s  CVc&[ LaQ'B ر6^tU-7_}J,YI_uq8uܒbO* s>Nx _-΃v6ؠ\ ka:~S1UO-Cu?)?qP#W` QJ{j9DJAhѬpNZaD)V1nhA!\, ʁ,;`v\CL'(8D [N$8ɴw].v,vF8bwZZR.B@p Ѕ*PBV`Qh ߽A!F.k٬8s]ee`2qQr={?`&}SUY?K]߿ɪl̞gA׿{,;Ͳjj9AZۏƜء}8 `ƠPq$%>v1ohDlݽ/lE'<$0eOUt+bz!pk_J)pui)jl.x?/1__:&`h-tU "ʫD%O&Frzwq2-ZuFE2ٝ2I3Iޙs"=k_Kg֚+a, 0ĶЈ#M# a7l 5Գx!O1C67+^@r0k0RYj (΃\H'N!GGfGcgM\!G_ ! J 8cHa0q @bk`uj궸.˷$.S}9JH3!2@7H"v p ZE : a,U֫Z0'c+1Z hG"V8!%Xb`3mU!p$@]ұ'c |tx]i->o+ٮIvfNLvݤzBΤV*csNa:Y(+-eN)bRSFcttǎЍ+~{}^ݺlDcYABI cрb_ EX,rq/5+ѺcqR!]]or+zLY,U$@y}YcAHW# 俧8#%Y/SH${z9&\A_,8)XamTN2^ay@iBG;{.Eј4y;EEDMWq|ІuG߈]E.3:z!tБN &5|CD2̤l?O8(F6@ԤEA=nxG緇ң*HO=8=mskpu8]-RgwvLO Y\} ϿWϺ"H]g.\N go:Fv=ͳ>'.v羣~ܮ{<шq<*-g|]qnm=hB{ p}H념Tzuٟ`6ND&Є5);HSe[eP=Sf Z;ZJNմaZI͕\gLBijp? Lֻق3 G{V?075UXeՉ [Ms`k.U y)HM_PJ27퍸/f<|_=ek(MﳫJW^1r.)MriJ$1 [2כs/)O_"UB5` 4TK4*4PRPP&!SkcS^M̱"!@UΕ2zJb'jBtIs?6t"(ݐJ]_N|G0ܣ{ ͯ/GTg )iņ cn`(ddRdo#UbsĤ̍MȬ4>i0gc)8X̮TCCH1RaH%\-aT=*6.MGgQt`ՙn lW5,I - "FJЈ\N%`CɒbtYFoE/ KU ƛؒң*^ %rڌZr:B`ZàV9_{q-TFTE! 1.T֬G6llk2'.I\V@ פ%cʥ~߇=Ny0}ɕXg2PzF*.x< Ng#xDt7+032q8"H2`3{Ho#Ku)OO^|7Fr3ݐ7WyOXJgHB ,>|Z-,u'~8]NC[ !NOXIL,68Wz;'zɆږ6nv]n4J^(L-"_D$tToX/5'Bh8@B\$&C)5Ïǡ4o(n&Z)*1?J[c SCT 0zh_,vا r\9ݒrqs zڇ4f7` ( Fd\fgJAi1jZVb!ks Z-};3a2؆'wD&O'e|ASz˔kcbƁƠ19`.ܬkAb+ ƪ߀ e\d.KI9Şb֌f]x*Ksل] \NG.{0]`I'S y؜(eAc{5h|V؎4h"+ ՀƥFN_{_)bKŬJyG6d/j[ iF qgF$ KY N~>LNďNߢ':=?%2y~//R՜.E{Go_Phʔ0LѢPtt2zyUXi&LƖVA]/V{t[1^GVZۢA[ CiC&P{|^L"CʢQmTST87RŔ 9Æs=\˲0E&OhZD4E-]1nd'+* y @)fp! ,"*7uu;CaPR)ʤKJ7KO{l S~t\&%On]]}UG]U3.$2Ab!\'rT= )DF(0簋3H:jq=|3~%5;𭾬{ s{|@Bя8G?].OVϦX)ޜ[aVt )s <]#Q#Ecȑ97@O Hz*lвb; Edvٛvl? n X,D_m KCY0Wj($L=*zpD}%fr?7 k_9|wa"9-M.1F7_^O׭{z`6;WǍW_?7ZVHhׁAckDŽ^`¬JCen,9vs3u}Xc7lEQocr@=1bsrb0@hd[N><ѦZ-Q']&.)s  lZNc5$&B;!k/ˬ,ϡ,J|{X\_.l74Iolz2z#Zh9-A"bC0ÿ!׬O]TnAo}^{r̎r1Ό// IӳUɟ5l-SNz72jpR0[1w5x7i{YG=X/n## &5Z-&+,_IPxl51_x8Ĺop;sXa n4IӕCi6&"s.[9L!=19\Qj=L&^++[;xB3zņѺ牆HOzRZbYBVx+O\MB:lix,ps Ď9,+@G\,%ۚ-@w=pb t}kNwz|\bΣvQEDJ5hT@!u`k2gOق>φ.Wd]ߘӪWżl_a.~wum#))E/LڨQ*B w ^ ܋=V7!g.԰ɷ+RΌHWA*UgMPRش^XIͱ,/"sѾE))==mE$9eKbZ 16 B$ fEHO1 C©;֌7LW -}zbq`C0M2vx@ZVfn= z;T6A ݝ?Vݍs\DU]iurNn.?-eCo.{ir0],ze|X{(nq|ٿ] [R~_td_zqЮlN(m\>|oo u-iO^~w`;Yi+שʟNq@=^_zVvKT~'ui>?io:|}]@tGQGۣVG;CjV.%|nkIsӣ0_G;.{3m3 )FVВ23`u?dۺrV!/u%pX<]KԄjZlS1T{%}Tw 7įÕ/Ns7E:AkɳiԴV>G]i>j>c-,vj]']EyƐ0?{۶"Xɶ"X`o{-mF\I\3|Ȳ,JLیp8$3 f}e[5Ƭ4SJii /n:,g"U/)a!LNN~I6_dZ~+ڶ#O}+B6 m-a؃d4!8(Nj,"(0r!Ʀ8@6 dKG6s?6J$ C 8 }*) mJ " nr7&^ t6'~? I7)%4^h;h^'{t>Ox{1<xl\|9-r͙A6ƒE ,C5n5s{y ' h4.Rk-4IqԎq*tTR,Gq6j<[@ZGEdQ09lVJHpD4QgVP[ՓТ6x>y>ߞY=U>Q©GL5u_bwX[AXž}>rG#)%ɋh\s?d3Sx;cv)2ϋfCJɮ "zo`ro٤,wo .1]֭(JӾIeҶm@lit~\nJ]x- |9MF﫟emIgG~,&Yͪ:btr 1j {t/{ 99D,(D>YWR?! ]/X%]]r OU̧o)fz{,˗  3:7\RF]C~?kyǿ7YL5=NyW{uj;|AeW$p;pq6E}ZOx 킑 I>1L,KH2s&(;h1+|:O i0<XaT>.'+Ը=LNfҾe}<@NfCm'Q`JӈQS<)*[N(8ZXf|(׽k^qt>]փ}U@Y]w't3b~DkޣQ-%Ҟc9e>-r6pD0%S=FD!Wcv(6˥ӘG"h0eF͝QFGS@$^1 Tѡ Xv6g=9N&東8׋·φ b{m{{ Jv,'Ed?K2Bv#5/-x= bHo J#_(d}摫. 6ˮF(Z_0YBdϠ{ެLk@FKQDf+ė 3S(0417?Lv% F>#RnP663O7;! Դ^-k~-_:<>벌 [#A3Gȁa?24mХFHs}R bP /E1X%;,D!&(-8:FC^"8qVQvd =!"_jkKI DpCRi,-,90l\.b gApQ{J/gniUZl&I{h^_`M-S\`t+Rn C8DXkRIY ;PAZ%}@gk)Atx&TO4ϳB`' M@QiL`v1J$#l|,y`>hN4OƍӼ?Yt #k_ꕋAA2*=2"øՄ#D6:GKy r`1w&=L_#/&h]G!m(ٍ6wdGx %nt 8&S?s5abdԖ1|g:z*Bx TH""$0 {A#پ5M w~\y`>d?'?ćߟ4a4 WɯOt2 gK3^ɌU_"wЏ;ݥw]һPz'H} IW$m]MtNWuȐxJ[NNs GJlY:Mku 5\'eBI.RI]nhᶎȵP02uĵ;J^~m=d+vwӔwG`5mϞ5: -BvY ޠ PJObe, j=6׻80iqi/TI!zbVLa6}Ù?n_ݏj좻aD5o;4+f[ÊҀ<$Wr;*^-n}z/B޷tmn[VY5MԬ"QB /XvqӭeAta,7Zn'׬qtry%o>+=, )'S_$(P1Zx̥*RUN 鳞y 1ѶY[/@;0!dtq% ,  Q`cQFbĚuI*4]<{nWNO7| >Nͱ+D)L߶' O)b"kb~>6E%-c.O]DO;"e!ά2 %E`Z4DO?L4Sfo{=bk,GB`|Uq &R%lPR!UX8cG `ħ{ < ƻ\8??S(L%(ANY˭T8r[11 4Kϼz ]VLU> 31w*/:b;O[D5U螻84jTک T+bEw9 VSxݢxOyݢ8Wx"u{ݦHa4'cZЩZQp꽤FcP+˹dݦEQ{@x!Ieꥍ8F?3x`2n/׽],J -UC)e=m2`MPʚc `YRqg&W:hM">ò%_Uk)\HKa)@bR'Sx1JόbT)Ԙ1h鿈y1ŋ0>Nз|t&o<4H*>ljcBBM#(*֦5k] Ѡ ԝ V8>fm]!˧.UWu Z!u]{8t27he(JA;j@( RG"2Rit ,3@ ;Mނzcsݿ0]RD) ?%baAi#6YiuDfzst ?,Lm6@3cGdTa`,ȴ]"lHGc&"X[_[ʉʏ1eQ/ړ+i$Ere6 8q[8zqm8д3bFiq sb~-m--7v@ZOT>ҺR_Nn;[!rFAo `hn)1H4POkR:9*j&&px[P2*|By<#ui! @Љex(rl(n|֣yMVs-AqΕtREF+&l l=q}7i"w_2nJboЛ5v7Oo$IV)(krm̽x! yLj%uϱW N51j C24QbϣJ ,q!HZ鄒1 &4i6(҄HBr;qzAHs %'Jډ%/#oqX~d|3&/ENՖ # K6Q!@uc/aڨV48?Hngۤgrfr0.w+Ƹ婘0VI!% Qgzȕ_cX"Ye $5ʎGZ>@{%Y#ǔ5ݦ{z3cE:;iT+0USP"hc}ttXVbٰlgkf7Su_2I)R``)ǂؘmgK9"lu Ь#UE vs*q&W,GA @,TlXS`bN A|y̓c2ܷmhKj *TԿsJOUTJpabs U Tg1pUa2Sj^=.~oLBSV$Qj8ȑG (E6XћDѱ]שo^ ]+MD!kZuCsdB95[L3,zrN>brfF#MӾh0 U;pv z k_^HO֧ܶ^~K?D? HM^)Ē5bjVzFNzi:[\,'ver2_*0U6Q62%dj0e*Q?T 㗂[H\Sp5[Kt{6BMRTb&1`[$.~)q?@ פ%cʥ~W?{aӽ+y=mwXgJP^*y<2Mget7)8)-GSf E7#]R]o ŌEul?EU}OlXJgHJhBߢRw%֏"|lP DJV8̕^$E~߈^e|F/ ^E䛟׃Hl,g5ffKKx5l8f!ꦲP7Zl-zvȘ91BpJ|jzo{hBXb llG6$G.#qKy%jiOo_{\1@;SW0G/Ea~M\!鸘zc`5-a10@dvr}bSOL/oɼ 6<#2~4!5[51,)(ƢcjB4pId}0`w pbZK"䘒fJ(.M88up"O=_v{鼜$B1an^u6;( F)*jev^M0A+ۑL6>W hklo/e!WJ$>pݑfpK?>_7"QXZ{\p\۟arSt72:PzmCo?őks JWs %qhVIR)\/Jj<:mpVAOxZ}ZYYIY|曮&.Ů bZ Nf5\L0Vvj+ͣs!^YY^BY8?] ?w&l66*0lZm +8!eCK!,b|$?ÿ!׬O]TAo}~{r̎Wr̉AuvurA޹F<`һ)S&3N 3{ނw~8?wEG! usx>Ⱥb$%Xj/ 0U ^&++qmvx9k,]~B/'Mzt礽Gh؀)]J=yS%l>c1 CpZ {M6tbe!IQ\A,+LII2Nm%1oe=SnذD]E]*rEJZӳymա >.֚ۺ܀~N6<{W|N} \&45dSJub2κ0IhQ!t|γK~Y\ 9+]n[Jorh"5rY8ՈUk@ j^ k*ey;qeJmfd̔3* Zb9 QLYtB Pf#-HH&2h-ӚwM "T]R潋9j8+XE) (dlݱf$ Eap."P&}&nC ҥx)﷞=*wn_p?NP|o9YzQ.ڴ:9^v'gI|Og^tCG `v9rK#;֋vesLuB֯8_?C]too4] gkWEp`7NT˓Er`˞}/_z^vS'u:u4_tu:|}'GQGVGCjVn>5|nkIsӣ8G;X}?~#+Ňt{ k%`a\ǴtЕa]t}7{S#yjYlLMJ.dn[>:?_%'ߧƵ)?Js׫濻jS']uqƈɻ_OnY]{+//>tr^wwtd7}pT+z'C/XB87 ;=ܴ}PoEU{fHF/k,+UB `CA.ΘFɵ1 6Cmn ˴֞#8k5x&!]Ht.X7 y yۏّciX`x(% y(1'goygN0 V)NMRL8?&v\vo;{ȸlװ+"^QgrKm6X6 s"}Aom,8R =~:f+ӜKEo<)F!0snRo؆Ѱ<Z=*nu-_ -6ғ7>T2.մTp;:h n=5y|+ApS)v3X&H04IPB2i:N[N3B;e< )Sɡ\ab`j[oVjR %jĥ(p ,] G0p^yVjw%,w(b}1CWVg<ļpt:7 GUf*؉)w6Y^rौXKrV(I]Eg L eKةnwG5dÉAj-ԊS ͔"mbKQl|H8ܰ P8,p&g{HWin`wu8 ̌b)1HI =M"eӀ%ݬẐ gѠ^V|+ 7z)zV͘Rw⧻"p ѸJ'iN.7WP1& tjtӃOwŽ"bcy %ӆ\hT[ftgy,֯c~)~0cGpUM %IPMM|>rh[7B> (}]u 7f>*jQ6'HgUhh/iȓS~JӒraX.)1ePnvﱳP4:ˠU=Szh&©SiC_F\?BdvIn4"$fi m؇.^,yN.twYݽ}xtǠ=+l<༺FMݶF^|`՟{#/WJhx#]knC .zW.r+ې|[oKzDZVgoۓ~KOjiS(Ӗ},n΍nA4IV/7!0e5+`+H3oKA#.Sn +((dQ@,߽ oM{Eš v~I:ы6c f҇~J n5YYkB)0pf\I<k}ÃKO0`Knn`ӧ*cQ[{#_ڼ6Q$wߋo.7#٤89fp%OP+i 0UR}?B/)U{64 ߲ˋ>G?3pKۚnWT[m+"l_ě f<[|)EO?UWPVZtFɟ A(Uwetzxӈ`V}p5ON B9c]"/<e,!]Znu OۉA4gS2㺼Tk3,6+׏lCK:>P만wyc[+>LT<|v zt7eAR]+oΡ-FWrBi!ZͲR+te忛>?p]. =lNF6X`+BaHL m0˒к)ގ] ts2)mqi9X}_49Ex=.Wƛunɓ(GQ)ic^r쬅 dRKJ1ΊꌎVR& 1A}pJUTsc%bDR0+.{J)=ٮ f^RbidH`"50$ ^ gDbϣJ ,J'0PҠmPVTuGtOȅԯ_gtv$CO<ݿR~ "vfF^jiܥ2|#ԋMv=/R\j'ZaaF<:˱@6jUL"7Ο93-W` hIJGY;c@*)D#J"h|fP4T 0rF$Ȉ&*FT8G"B &(2vgM%eWfPG. ^j5|-_ r?uBh< {J2J1ih(XMv_VgPg;l'5{|x؟ӻ JGoFh0Z;(WSaHd_3{&p \ 9GN0v6R1HiX* v k*wtF!r;ȷM}vS`sD̷4NZtcJ)ep6y`%>ָ\uȲG8\VGţ4B)%Ɨ6 V2GGxwxiQ7Bq'1Б( B) V`>R8c4GԣE?te)]qn{Z<|.KpKCXiuEǭ ($PVTE(ڑXa3x;arYvFexFĝMlt=ŕV')3[RLmYYR )%1KH,ô6le=+KOke9oBBqA)y<7T(.qX)':v*@XG" (.X TxxDI捓Q48"` (O ;9$&X~{ug`pZoZ?u(sHW!(xH g& K[q3SujR #q8rOJPMZ(FfQBdT,p MnA,3#uH{ZH26ɱ8J0̓Iœ`0Ij.5c[j[,},pr7 w[5Gx;MnpjYW4 b9_g6Ia@l1x6J#\j ç=+GMR`%h~ }>|Mf㪆GY-em P|\ ڄz c;2ZVn<^@"{smy(f8yD{ʩC$*\HFa@vf{nJg:<+~4^<c8lv;nE4S2UR3/P"N8EjB ,C#Lk 6L>_LgIo/@o%P0 Stb !† wjz}DlP<,U嗉i&k|G&[f^oA~;oy.C\+rʭr$#Xt֌+RIpLy&_.)o4hR~nz[2 vz !oh xL%{IqP['XmJ#Q\aB:cZl\օSllå+FY%}]HWgDGiбTJ+kވrЏ%O}4إ낥39N3N9T?^.tbAБڠTAmv!Rx`y0#R`;2(:o;c,={-#chnE*q10l:-[Ewm$1.AN.GJggZǷdMìRbw,^*n ZذLuxQ*n,&h"S Wؒ} iNe"Ҷ A2_Y_*ˌXIF{T2m𹱢6cf)3,e|V-SQ!oF%e,wZ|OTٴ>y|=Qk ^R=/2>V\ܦV6RS$tB)ڻ r2PCT\ShR(%MYRTlz0+b+bQJþ֖;-c;Ү}T+8ƶЎzmۊ? ',?/z9| x}=t=]s+ANfk~ /Fdo)vbMCmoCubQBf]\f49bN@ BDviD+v[4.Xfqv`w31WYP,h00 E0Q@&4=$c]̩`FǶV2pĚ,b*dT6cg=.ӂ0MVqE"-hwIu>KhN|d|Q^A>/6M-bR}c4e [;K1jgEf-Z4ٓVtdw3['6cgEteQsu6cnlhG; Q mN+RYrįbgH..O#8Bc{x LXĆ⇁0~ȃ1{$OH'@6~췒#idLt?rY~khWf$.bP#AciAy8|GuF,mr!$R`<)_ȑ JE CftdDQ?[m*Q|T!KLJb"5n~x}N$yRݭ!<B15e1o0D3F(;Ui ig +:x0,M^u$Ҁ!EͿ8ypw$/汨@FKb$-Y+:gdEO4b& *Y,RdLcWv=|ZEv#\It$jSaDIl" DA*FGӶ3I\DI(BBXAiλj+YhMHyAHPmH -QrΩ0F$c#5tձ cn(5sP 3Hܮ{''nsg/gb~J2J)eCIUI4/G 1vߋ` Q*0.P!H :G; )WoX]c9d&E5:I`Hf f01^(ǐ%EQ_䁸 yLQĚ䯷+VH^J?~g?٘nrz!E7uNky~ǏX|4rG?rF.Cݯ^NMPmFHL:tT/LdmYp(Sw[tͦ/S؆W_Wwջ-_g4=A^mdwg,OZk\@w~;2άjr]V^zy77,_+bm;8otƧ~kȝ1+j> 7/%ޱdy+YWM3o5μd8fTuЛ&F8|4Cd$DdQe۔R)FY@*>BFIA(~S _tɐLɧsZs@Ay]g4Z$nyog3.4/M7W>}wws;Qb{ԟbwl$.qVjQ@Zb uJd#2: o\3MOsYri k#?N~~QpO=l`8Gu z^2t>xG(0<6 w?li6BSy8e{5~Txe @rlWqgֽ|=``}o#z'epp2\nwԢl;/VR(ǯ021 a(_rEe&зh ),)b-v& Z_ 뜘xt@R!"KuL CqN+heMTZ$ `mAḓO? 6q1?O]q`\g~7Žq~g8o=ZJ[EZiLd1edc=OJՙ*GV YDy7; o {dOVh *>ro^NoM|{F#?y~}X-x8zQ3jbWvGc=1sL<4O6Q%4ɬ5ZZM2:[-0 Vh?T.ւT; HZT0..h`T`Uk Bc،Z-R;}N'4B!H:EUy!X|CqrmNsO1X R"cݗ} P&Py(Sʰ]TE ٯ/@k%& կ1K{W>ź*pد=w,p}Ǎ-c00s-Ƞ];P:]IJq#~$쥭|X{ڌ7_4&0z Qge Rz@#DS+$K&+ya |ߔ t@V?iƒf]ۃ0=mjp}ť|o:NkKh*1" IЌ>/jK"!BJhZZ*1>'`GG1 IЊ%u&Kk")1d A'<PJ&)Th^~_-/+Wo7QiI:f 1B"^K+F'(!'d6Awu\67u`z3_hzMhy;_HI% D9 "%CbT?jSX cr:n <3l9},YY=R1% L)si;#1CG P>^Z{vӴmÙlZEI߹7~ J&P Q۫AAB)h_oƏ99gGvު`tMJqBUu6 _qVj?7埧 Tzfl0 ,w0!-T82/OC8=/QW8>(n?'$pqHJwMAW܁~ i 4~B2a&7Z~P̫s k,j~sih/J aCfӉ[:HFwH0.$= nMs.pXrIG,9ķ8~gDᐢH Hj9ݝٳדp?~5~pŮ)vŔ׏^S!{vH+056PB_0jw^\'.8>N\0}jq"$~@ʵc"Mb'BkRNh*\Dp[.ڂb$=kcN -FM1LgXMb >tCs cbl\&bRN5TS.)^N(%yEy/7`adz xM&`3 A>_dY83mqov2Fׂ$FIr`8y7~mP7zx\GGGk$ho->V.˵'gŲ| Ÿ.\Wy36jOx˸؋|I _XDB.8! ~;v淂^p >{` \ɇ;( &)Pd ɇ\8ɇf֌9\Re'SpǓqx=(|+j0atÏMDi50rfb#RFW#F.~kN2/((tn'„"py ɲ&DV7dŽ78 A8T3Va͆բS8֩x {;u;+-$ًŪhnYO:ȏnWϟ^x0 ҐHtNH&NfGI܄e9ºcCl}MXm%q&$hʆZrJW<7#%=\RF G!3Z4t@Mjc˜\,0PJ[E7l¶h|9760Иtل7hŐDΈ 1':dH*#*"CGN鹸vbl|qRQΚ^qmT>ys3CP?Jjؑ849XܑX Ei,wճ&veNjbRڢ7MRwwϬR^O &@z;#E׼簟mN'7$ Ʀ7([ cq6)/$5`CHN]돳5 N6*maH,LfYGm핃r]ͦ;_.8Nڼ'_KK 3͗<#ϞM̏Da˒ 5Q-Oz,S>˞;\B_;(5Q;v'b x[,d!!.pTmc;xv ,l fuzOnnᄹMx}}LmO[͜H}b |j<|7O\' ?+P.€scCqj)b$peBeKAyPu 'a洤z{~nhbl):Mf)w+"*DͲ,iY+"@Z߲:O k"^iS[4) B/R7EE2>b"ݦ#|c(3p @ "8@-.,ՊH.h=XS9 \K$dVCE=gw`Y8~vˆx(FLUV6+ "x[46_Ejr&I cZ5&Gmr{Na6 f>&#$J=?3_HBk%1mcr&( Gu;-*Nnhz(AM'5FOZܽ^mLIZ^c[:Tݵjkkyt/jT]HԴE vICӋbܦr24E-QKǴE3HA E%j'}Y<{)˖ $Nۗ Q@j/+/mc 05b(.W~us-XT#2 9_b;e  FbSBEBmPF"]>tu>j;b|6}T;+^;r˄#)4۝w qx K|coGǙzkFmYŎ ߋ_ #b!)r؆ d+5Bl1% +]_ΦK(pp̗VJyb=LnS jr+Oo;֠wca0RBWgj0?i"$m 8cLT]Z1"LĹv3..ZD#;D%9%j抪n5:ee9cJİbZS;lBÐrd$ #aZD+_t6jQVu{勪[Ye{@S~Mz JcdE{?zs~KBUҎ^Od 6ad?-UGJ B`,]Q<RNN:R;] ]$@v2VSG!E^ybYԹ4DZviuGd5=* {JNg. t.?6 VC^-I%v.I9V|/~F9cEބ'VӊƇ]"oz ԧ qSK'B"~.]-8FgXT@ uٸ1t(;3q~j-"nyo1;R 0:|ޠn^~ 2rDxMѰ_sxP>‚jnsmkH0H=f-ecXZ+FkZ#~b7^i6i2"E,mx^جM׉˂iBVkit{.T+LB{HDT;jJD!@r}Nv\N%x\ƎnɻmiQ%i^^ҋwH -ٽ0-`Q#6@tsq<(!- ><M[9_J#VxBIHl1‚aEt+ՑhP.d1Tڕl"$z t[ºtڰMJ q$6Of]ohddjd2k> !Z{0"X.]nR(q0{ˆq)ܕB=|ZMA3JlR4DOj2>ΜM]-0{FX 5AO#v 7PlOp :#u '[!8}g;4RBMWS(֊qq`n@&)\8,fXFzqDHr%a-a^{ZPʏ{wD)Bã0!VOu67_ zdE$, r?iuse#)#7ƥCX$6## !G'>H">2ʔT*9IGȌC*(9gv[Gڗs1HE`s3 ڲ^>Eɲ(2FXaϯ8šXbL& ݥz;ќ C1BrM ( @eݫΫ\K^fFM-Qj%DY-om_U:}rg6J;)ٳQo)-Nt^Kr/$%bP(icF/mYcLcvN43zD%)p M%L |gQ|x6[U̘o[䦬`~Џ|2LuGXlseV6 ԡÇs^|, FΤ5^0`цuQ:^U$UӦ.pCp[γa, bFת)7^':Ce^N尴XqE&vɉaet? kZP[> !S]+cD`c\f@& H'Yt"ʹs .Ẋ6ӊAHOv.CP(b L*0N^!2CF|#CW ??-5LvBfebCIj(Je:Jd3NzLRW54#aŃ9pF|C2[40rߊ9Fu c釗P"lTSڲXbz$P/>ɐķqQ:b>T.|J͔6,$}(1%sWCSk/TRdU&;h2[ V* K Nu4LM'j@g) N:c^_(9uC{Gw=H5)8DE*խڪP[(:)6Fq67B{ vCj`E={jM.N Mf9qTYRHU߃KR^ڐ@bQS@ DCMjM@aE<9w{Sse; vnf[4$1׷2y|!I=;@ř[Jc9jwJ0rj-H@Hї[<{݁P~4d[ײlU`z͇]vS,f1,m,\NHsH6Zxd?* q] .989 z8B3}P㼈Pʻb45Qr%(z|̤aC*[՞C0#{tGC_j8iĭb F>qɷ))o0_gs!$Twk=&B2"4ǡ: r :^Eh ah*oqVOB@Nd- M0%CT_F Ӫ2MGoDhU,ZZetTPP[Oފ=u;{'EzdLZaN7 ~l )D%ݎ1%*/">snW8";dѣ{ژa{_]mNWIogE%@5Z@YɴZfiVzD|8g1OSExkeee ]+c*T RWYQ}6K*+b.)!qJ{vxDZAI@roͤ6 ^*+)b({R+ĔO_֏;8N#v,./2h\y~8Ma+9^ڽb,Q<5VU0oŗmk "fJ)H !,=%:"Ԗh iO*eq.sU kZX "%J;xfAZEȨevVQtaTtMm$i0 T ^f-k38QV IzT KDY^[\w-U΃ϥrgExClɑ6HÎ;%fLR8{kYTRXJiyR~3Yw'W+n%t˥2Cc{X6U:/Z@(@%֬'6sI9S&i.{G'%ZZ>zm^#"jRCF+Y)T@Q!q _i:.B1{r`ꡘ3Y׾2<0SojI&oN›3֕͐)x{Z' ]0$WnoH '1XNYuQp2aGłT۬K\:{lFtn)UFm_n,ԣ\wHWr/ ƥ %FNALy6҃:zS3.KVo*}A:!Pq ÿi%9_ܼh{nmX`-Q#yN UEw`̝*? @XQc+}|e]pf%r9z.x`Uœݾ$<-.=g&rr/H!.q;K@%L$ 9xy->6S]mvw>[G>ޒTr$ϒdrS2.j}9T ^ErOhܷ͆4Y u>V†|֚g%ڊ+߫fݢQ*pЬ3 1&`er~'Җa_n (Pyr?ZՔK}z`̦+< DߊX`qR_O>mTzt|CA󙁔;mCױR]nC:!ʥ('Лul65+jzzJA+"x#TJIұ75Rk IA|ĜbXƛP 3}VEih1]T#}VТspR|c+ilB|)"Zsi%n͖Ut-aݝ'xN)R2C7Da[0`y{1V`Nnk t'mtGApiO 8 =ad1DՎQ) yoq5UL0 $f̷̟KIsFeU5(ޜ_U8rܛCH Ya(QOұ0Ydԁ .Y`[&2DBںs\UɳG4Lewm=V lU_,"/ }P]euUu{[lɦLIv5qC# (0H ]bA+} XPO",<{M >i}l<)H:aEXJQnaZ 3Vt_b7 %LcEؤ~ۑ>Ս96҇_] kMZnÕξ޾ɳ~'\I5ɸ"Xu> #5~P`pz HfQ mT۱E Dlr\P9x&UdRώ#PRk4N&TN'p LhL*v )hQ]Ge+j_G""aLn -MNϬPW+fgkY2{ga_TJ($b сnYؼQ/I]~Z RtoWylH(^ZRxpdL6 gmڡ7̴v 4O$c%# 7 ]xqHÞ>S7Z ;ck";=mi3$࠱i+F r?m#: 6Ju;S".j x>#k]{Ы'3g/EU݂#b6__٘U._6[ Xފ{˅:i<JLع%ѓ?yyQtV*x,JaEΰnz*@@n"{1߻WGrʍqeIEa: mm }yEЗx̏M,T ǧ7/@4O K.[+)YN(r; 6-HdGPzF _}X471,&T^_iw3󜻣WCzR4HG5_B=I#'c#\CcpI讁z<6)f5jŤŝ;N%*jh X6Ĉ7\Nn54"^E QUoѢT8VkK֧4ĪpxU)ק#TXc&ǒQ]’aD^)Dž0 蔹Ǹ̗kE;x,KϸC2]'״Yh<-O+}8Ų SZ}nI5l_-jxP|"AKY#YC8ﷇ}8Ů'جvS _窄h2e0㟰6KY֔V>Kqr(a{JV ieeD[αME13T C1A\! )[+#Hb%ITtm3K\DžX%i!A'lO)qUiFك6 :Q%1l)2F&L8Z.b*i8*d*1lX4 /eëh;ehD1Hd5EJ$-F;֦|}`6X3ykC9CNW2YE`ד1{xE*J6&l3PEiHL,i\\W;ZYY%]tM=l G=ą)sqDM:;]ü VO֝GAZC*_U1I/[иFMiwurNm +~[ +meq;-uu=*U||gmܲ! >5?5ĊM*1^1Z)ʮj B~L;&’/=wI$KAB ;@9IbqB}T_%GYojӝ@a='C=9/偯RvщV=߷{8R- (4mY, ?-wݿW9]9+0dE?-mfplTg6pCZt X0X$ZG !#Qh2LL ⇛cͲA9i?B[ϫ|<ߏ>(;d}7AY6/ȓma>H"$8e2+McXS0Kl> ަ4 W+ykK^HZ5~H +F{O&4xx$GLD;u*B$Z F oH,CaWp362LDfz$6EVK9=K}kFurmv/HW}vx{+|u?H3Y>V}oN*1PP؀{F+fzVu)S{P|hf%4F {10=jtOW4,vDjшʱIKHЫNGY3]bGa)T!E6nV`ݛ*b&1Z!-65E,)NJH*TGapkR҅`KyuwPy1=޲G;) Ս60ztNdsxE,Ea\nA!uaLeؾt :8HW'47 Ze+$:rēBAf؊( 9`Y1~E띶T%\W~L+tPVݡUIZNS$-X04 E}p&1%`2sX0<ҧ >m G!$q`ӤGD!d,do1Fo f_ j97&Ӵl#vU}V~#}ZIlt7=4,#Asr=&ItL 5IQђC׈G 0)v>'O~(.T+1e7݃fhn^ bB BD S1F*ԫ݀);g \<%%\N_P;J> voO AF$L@NnDDP@ Hm4v1,7v-t qCcp=&[ljzƢMog)}Bl^oe rGa#z+JR$SqJ#i^"-"qVH'>I IHk|\` @.÷>yd.ܫ/9YcۜzuK$T?0ԀKbXDB2ˆEm0Z^ '.^ d^6n:Z z739=&A İB ST.)\Cc<-FhY044U@q2%tnaP⎒`o7ec| 6* \ĝ/*ң9k cBZҸ2 JT!I*+|&*L[Yvq#p[O%jbb"cDxmFS0i4I]Վ!y([FQd3/6Q&&Ʒí٩>k~6B QR2MIBB@Nix0ibDnǶc؟0h8v?)_\p, }:קvU 1| :?6Mh~N'{G 9Ӓf7)4@ėlt 41Ce%l4;xzvc0)*`}U, IucdE %V##" 5lLWe$e BA~ 2reQozŸ_<4'vRkh\bm|DyϨ@Fbeg">`Dc ژVeD?%6PޥHH`(lX2z|i9 = k}xj9roAp)Lv v |=)x 3z>ƛ~ǺugR^=rI=|>:%K.F!hj6!Id1"DuD^}eOA1JK#(.4k`ҹufk@pWogѳ's zX,5(+*I#jhloF\Fs+42^5]4#nkI{HDƐLokJ )eO[_"Gܲ+Tܧ3hɒ]?J蜘$ NA I 4qF) $QTkh :S1UYQ Nu؎G- bԓ\`:% )Uω]-;IcDng,Rp՚+8鳺nhwT:z+xE˽ y̩c 䨸8:~a!b𡼊^5I;DRjrF:եmN+gRڃZ(I]Cqoaˊyl@?- Q>{Ӗ#vf(襁ĢF@ %4\ \Q+0=YEICPh15@ E.!Q\g쳤tetciZjDB:0;`KS;UVw*yW+Ꝓ~=x=h;erS4@WWHGvjTa"h6ˋsgI{Y_K)~F%Cdji0oR<\CcVbgP"{l2uahzF "]zVp,f>Qi r!Y0`24V:-d+Z)$=b )6ˑcarJzq 0E"f¹(PdoݿW-!c1ZQ*?1Ba<,]B`n!=6Cl5T [Щ: -Fc FeȢPȗϻ $7fld`2 Nq4(a^vEpfaZG !#QF F1XL/6'nY6M'c gO^܃wn:s5p`oݲvo*l|>H` [4 pWi` Z0Kf3Q tq:KԔߚ%Rs)n8!pھhouڿsBR⛻mB&[:WNWՠ\a<0ފ9%fXtmt?{W8 ᗝ(z fz[؇Ac%[WQ uH))2efeEFFD!Ap]] '}?l͂ܳ8]\U\vST8z)oe射Ͳn;-&8-f^S!ʋ0Fks3n)ϻO՜5gY] ls=W׭@m}mtsN% _UKlƷQj4ؚSV[]0~$L^Sҵ4\fߵ@eӄᴍvll AA b˭;<^E DZL`.tgŸjѠ"Ă#PIhڀG%f7ƦHa:Q< =S-UGϦl_Si8F>He0)h461?u$̽NGhvZq.DyIe6Jx>_ hy؃%R%HZqu$D~qYUᡖR'HoyԥiE V c97AAr7!6%7*( TCc+L;JZ%&D;zЭīؕ,ZHw 8 [y HU@Ti"9Kh^%&R\(@&| 1*/XhK,%ƄPJBku&qVzFB/3/Q0{,S/a/2,|o?xi%d! tWq6y6?rᥳEb(Jw~h}1/|r{'X A1KܽR M_ }(}Jx#A&55 _bVG|;b#yB xhlpBGRGz"°\ ΂Zu 8PY./aîPl3B9+~?Yts$-AOqUkivUHH=&I=t>>mڽp'cӚ]C[C< {Bpʐ}V[3lBAٴ~%"r g>Ĝcʂd/*jK۬\ۆƑ:hn2g 4x4A kun8:B:C&4}nH( di5iL;W r6|(3X;a{nMM0-xN.jWܕ\=tP743 ɉ}v411bPG4obkhl䁰W-T#@V㤫=&dہ']Ei4;vwT1ҩո4h24^>+&B21 )H(k>VIT(6TaJ9~uQٸq+a=gf i=n-b_z7t5I qBxH9+BJ/Z6<@Z!I0~=HZu: 0Jϥ 6ZX54ރ̈gBN$(54^aFPz.PnE=؃b FOځ7R7-.2. aT:cBH4vqŜ=h|jL,$CC@n ტKAٶzZ Ӈaa轜 mk=~+k` 6nw71$̓Ծ7c5fYCa_*QH J4 G`IK5l9xLyKa1j]n6S|&txM,8x #&u'$+-bij٥ Uظ[F<H&|\\mt J1HG{7`> b% 7|kՄB Ha648T MV)l;N~1!ɯҕSgځm/eEZWpL!^''AEjio*ZRg V:⍟!,TqZ#-emCΕvv/654zj \G-ԥjhb"aVAd5EFXD~SNl)ʨ ΔV{{s͎1{xsb^[ UU aghbt}Vukh2;K<["- &""VZUEaA/.p{-$/+qY)uEf gu;jGnhH.+vf]'pff.RN[# ͉KZB߬ʇ`]JgIiTEV`y^]ߍK98y ;(v+5l_@pw)w tRΪP@}ni=wqv?WVfEo̔ܐ8AE$6T s%6;io/.Ɗ._fgbn2rV=DϏEoǿe.Ts ퟢ:3 ƓϗޗbuɻܷWu;E>+>%rV="aޜ?x~Njmf{8S=|o;t` v$7zÚY-W/ܧ[ rBg~sL:.%j~^$S L5~{]q4,s\n0zj /!})T3M P-4n HMR5\|q-Jf:Feat fNBP?4(4\!ϝ'v PT5 7楝\CSng$YUPAnZ tY&w-81ۭײNBB퇠փB_Cؖ Lս{8 1 -B@aɲM'ާ<*y6㯣$rO|d1x PEUU9D9<rOPٵt!?O0 F %<RO˫0Vj{Aڵ{oSvzBTx*#叕Łc5a1%Xu/*s,GA,cN$c#.Bo"$8*rY5FY\L5BRi0JcCqaꭰk~i$:ެV1z9"URkBl['!"~vՃZhME >цkG̲m:9ZX2WDJ18( e+B Д@:xr=1 C[ɡُ.>T)iv5$ʳo<{4{]εrMjf~JqV920sV/hou0L}W+bDSV0mꮼ [NQ_ɹid='R ٮfaŤ* /ix%0HEp5u몡~w?R)nQ+:hձ'FT^XS,X UEXw}{(,M(gT!^v=ٸEh8\}nfo mߑa`UtӖU{ϕqE}ᇸ$BS!-qbRb^OQT P+E#ݶ?ŷuK"I`0z䒐eM..BCC(~_̲l SPl1~x`zW_}eUnm4i_f09?'W0N\ ~'$KK*ܗݝ$=|Vm jhM!Zde ?ﰇcZ(H|bBICy FEii;nA it=K5,fpzvșqH7Ѧo 7 h~[C]+]0=YuD1SG8侐 T~kp0f~ +mHR={fżxȳuPERDYl؉Bc3T\4^6K}FjĤtUpK+J#Hb{4]@٬lB)39 SBj&d!*2@6+%5IAΓi%S_'g/MI5k|0 ub׶5&7gT5k|; :WH~˄k,u1 Bo?EYcyz肂1H!O.ar_DMVoaBLΣ22HwHISjU|#kXⳍ.d8p!UZ-3e$5%NuيStׅ&7Tq 5k|sv]  1Y@" #×iIM̛LYͽyaY֨fuT-MLU#J6CWNPKQsGxAH(3,гPó[Q{r/NM-ʒsVGTT{OEP.YקrtwdVLV¢X[SyF8I^5k,N]w8z~MhJTٶemb>FɖxexhQp|R 8DjPib[FDie_#ϋ,x+\YJiΙWo.Iy5.= pjXCZը_k*nC Ry4`Y;FcS ]Wfhf'Vm@V =3,Ӯi{9ށFլAǰ0--ݟ&. % Boj}FUvﭼ5t%^U2O"\۫ne1Av:'&\\g&07>WxC_[PvLi呗zy;oI+Mn%҃W6[\t ب"UڋANnGΙG?#F}^g#t⎙t4J q=Tk.d-亾r76F2xP矦?}-`5h1{%tP|` 2f[P!"boYWPvUAz]_&j |Vx.Ug2-k`6[22Q_LJǦƅAsPe2;se*vZ̔/l.Jw_ \('7@Dwˋ! ~@Ұ͓_ .`qmQ%at~CF!& ,<+cf[#3"E^eVz,m>,^>3 ^ۅNzM{@8wh^?&5?c_JG-A"w V b%/.rnșnmcDCu[j9%EbmC$#TQb_Z 5?cwIo0~F8g%f*j 1҈Vp"y!IKY<_cA?V;ޚ,Z07,abA-1ׯ`wz+ %3kiy^Vg5tlx4m镰 :yf>F5O|=- Mbuh= -bvĨn6)+CDA6ٷO6'vss@F3;f w,;1k ӨcBYE+%v9Mjͪ0@GGS,moЛsTJzplF `P{}VFSFxmcMߧUZ ++!V]g4=VrߐC9.Q:np +P_Mw)tAtKti*=]΢bylc)MZ{ J6*!l:ZC0*̾/!Z*p_#2[F T繂CŮ Q @wW[alb8m.:(fpVo5ee0&OBLX̜EAhk9ʘAg'TDjPz=Y{%늿Xt={ ᙿX+)ļIv%U`&m چu#^C%kJx*kvpe$҄b$z. P8ADѠl,3؀sFP%dʬ4CFR'Tr,0O ov?!Vyn m]Եe*Q|l iOXbKe hښdH_RZgx\dE=LZ;D+ kE^KE#4ͼEjv?Wm ]#9u٠\}A_bR M`R ^S@(n[ʺrpETF"qLuQ+a0 `UM̤ MyQ<:,*2ǜEP|(k< /O6Le]JkcPJqȓFOۘ|(3l5:ǥOhL6$@3sgia{j Pq]{4"qtrڒgIR` ք@; 돚oc^':>ޯBhy$P$bQkIrH!Qh2P ꒷ܣ5ڡݓ*)K>0 ydqVAni]) W h A5Z*kF4RC+)#B7RF2B&1Qz-QhUfܨM(daqgV;"0!&{Sux6 TɇhtQ3)ZQ7;Lj.JEw3(?C5z)6<^Ck4"]:5O4&E\]^KZΟsHHsg1TB#g},x_&Ek2z] m4+ղ/"V+-(lwJLQ̈HqBH`fCG.4\egUFG+ <RZgQ8D(:[ԝilxn=DkG+c碏M<;ȨB|l$1eG3R.f (J$=.G7=Y82ptgcmJ xRF rJZk.wq`nHŗNz&UJ~r7>-Q?;?LcVrn G?&a[HNaK3Rky( 7?O[P$.[gr`N "ԩJK,>!/r͔'Wn Pn$:Tl>?1^#Ƽ載Co·!GvbvĹLNM;P!;5v&4RE5j]̉#4+çݏc ߀ L}!LKi ~\)EJF4tT~ǿٻnW ,ADhf28dyYfʷ㪲n qPȿݜ<=7GssSw8 XoW?![8O?_,ci<;hys(wo/ׯC͇k3xopޖcd3^g18vW6w\;#EwYe,R2|o+HI0ybMaݓ8X06 C0< }I /D0{S09XZE#`>1C-bjaۿܘ10>!MQ!l 4! `&An!oVmcgm;ݻ[ۭ#=JJ9?ojass)|bt637ܪ;j!ૢiBΩTJ^Ƴȹ~%VyKl.~y8~2Z3=/23r^3"MUڰ{cG{>=CiM j6*m  Sp-6k_6*Ӂ2ۖZ+{Bp*%XOѤ65d230n{ ʝ?Z}nP)W`xƞwmu+duocY}  ڧ}&2G6>ȂfdBdlI&}kP= gSv`)ԭd9#F6@(ub4 <M`膘6Ŭ(l'vWzˆBo9Bi A^^iܻsHU,i DgQ j7&|sXukgkM\E+N5rl`+Z3͆{uCbir1g9Tm55z"tP8^ @_WM[ʟ~޻ɦgl܌EߺVq6e##O'. 8DgVZa6HD|1Z#N{/9@4,u`PUVـoA 'c"tI,.|`(h U\ySRڂvy4vTB) c5Qn=i<ːڠ܃r?}87<(WܫAW?'){1}vFrTy _ͳcلM__iD5[i{Ae>XN'>8UϲY} %RZ~6Xs9ru5ϧ ҬUaY 9ˌ6HވL2*SKC(+i9\'NGo=줅:9iY5vx)@Pj9KXoe?~0ݐ) U, j{/je;ӓ#m&YPh(67 )LM٢F2v&0 ¤wtw ]q#;9zKç6t#EeL'܇[_S A(&CM<|F Yc!̐1q&ёՏnCHl7 6V}\'aA^<-Q`PG B/p8;2V`*ZC0gլAۀoDv8a wOH6?GƜ8D{Ƴ?y(4,ǃՃNyq?}|BؕWJov2ִF_>\O>x7Iv\x炀o歹0.[^\(\jn5(%}Ne<~"k1: :ѣb&sMρPY@%hbM-$֒,&8^WyjOɚ] Rx<.JV(P"-ܬL`Pa1^mE,i!mz~m 2rt-WK.ycɐG߽m,ֵfN [l*p-Ca+:rLo5&b$clLgs!$ Bb1t\JȆ嫺+@_7˖U-muŶ9jj!m_(/8&anfqXĂ%qt"y/m X iZ|cr2v4ڶ=޸x $ju@A}o82 IP2O8i'#{5Ul eV4 ЂNc'nfNQYˋ?V:39PQUc2WY`jCp dQsy1ɕMnn.L \Bά|71rSb]EJKPbM',wYwoDar\w:e6Y/ cwҠ>MIUH]e[]۪V}vP]s>睌*ۣڹ} CMS5juP±ʲC/j`g1n5NdKPH{^N\+qKdmEc &A|1gvk֍ؗJhP4gWX\62&@L5&cyVL Uj=.F I*!GWtiL X?˅P35O^8[!vPQܗ*Y+Ny^/ІrZDu!Jm,I;!-S887{`{#6&l;ͰhrKsB&*VBBAH+ʪsߓa)i}>zj܉8ŽFHym";r7DxG2{e XZȠ7*wG"Rv=cI r S!"eJm0d9PIgXfqݳ~za*9/V*xX1?'6@E7Fu=]qΊUlYal l[ !6!;JiܲHʕhESEF֪UnǏѭE-2L.Y۴g]>ӳ$ҪԔEvٲ E>Eg_@Sٗ^|7<{3_DVc?60:/VD~7"FVj܈o$tfo\0:פb<6a^``Cyg(Lr - sZj瑓i̚vxqb ܴ{EO VHgP_zh[3h{x nlm-$/Aw0\3 m]wM0cWނj|w3&v^ /vz'/Оg=!(凗!rnVfkqgs<} u,6 t;B!|7kZVo N܀L+q9Q-q,VIݺZu܀ʂ}= by]|?-{zo ,xEZz^Q&:=G7C?}^@Ko:8ozs Vla΁a֎6:C=l74"ϠfxH$[4A"F$xߖ6hmE!)"{t WO6e#)1(vuak^z9;Y_~'}kM㾏{򠯳E,~-b?dfմY%vu6}v07׸A3WqfAq}]mC+1Mw sh 7^GkҸ4H[axԎf" kv;E @zPG>y{!3{LmM@ [cĐoAӖᓧAq$i>z%e8X`XϏG[x ΡMפ<5yƙ^?Իzx7^>Jw+[zίtj}n(L8Ɋ;4 uqq(]*?d{ WvZzf;A9]IV OS?>kGEm>\`&W4rkd T"#Y@72h H$O]P+5-{ &\rSe[ANwaϞkjYܷ!D5*0 1*J)F8 r)fy*Fkm M1d1x͵]{(ݘGg;le0IN raRO{{m8+7\c_^~EyL>?׶/IڻA{ej͌HWzgZ"YuNUʭ4t?:lt9ѡ=(6C 4W?7Tjz,{u+Eߝp$f͎I`&r;o8]>)XV|?H޽9l0)'>tjUOc=>ý7ZKa P~ a09wk8|UVP8e%tg/.U{yve*ƆI_^B؎\>DH2"^F7dA9}A5`T_\54nutr=޸thfk,_6Fߗ3J߫/&nx>2u@a?Ri%g^*^9M]Z5Dx+~c>}4bD X2Xg겕cdawu!gtGKf*1_/A;dvd&?O^g[γ_]jZzXƳ#ZYmC3jiò#L3{x `h[$ZizUGhQxvxv^U}RcWh$tP\jn %l81 v7CbY .-='h謼lGWhBUh\Xvo nql[bD v7J}+Z}|qx?|O JGxg}#CkgǑ 5/`}Z#OK3ỉ_r}Ǔ}MRC930E#aQW>rVk/Wi{dv}M=Hrv6 kzu~+e a@7 [G+W5N]+ht$2n _3uVxoW˹Ձm(zȎ[Ա } xx7U?=uюi-e?4WuV|mm㨕d[PDڏ(t9J#)׮O[D7 _f%TtRl VH:{g:$o8BԳ}bXCx髣PJQ%8>ʒA_ZLWGƋ=24#8[t}7Pɐi]sS K˞u^Khu,}ޫ OKWnGQ7sDK(db1OWGI0S Ⱦqs3̳P֦{;P>:>hXVE ~XNGyNv2]Q75 @!XLԠGyx/ ,vpyoKFvj ;5NMݜriKVYw~{z-S!YMCɪ\6'sdW'Zڊ.]v:> .Fo5)E_q.ikP%"b56^p*נN^lxRW!\fq9V!@qŪ긥fb``cj'CN6Vudwax`F>R[ KFtgeC`d3hwjWDM&4ֽٞiVI˶DⶤEȚY,ބ'ĔHV4 ${^ޮCw}14gP:N.6dPAF%뺲H w)dZo%Kw9y(i)ji :r VkrJ'&-ZbwsO/C./7U5|<,Dڣɜ aһ.^\սrђZФ&jb~Nٸ;;o!׃قIa)go\c) C%VN|nӈ1[寧d XgO] l5^lˬУ% ߪA kA,Ё3 sI~L^\Vm"gLmlz#c2X{gP{Vd.\jqO#{|"icx|(R#PFHMEs4H4T/R"6ccxM(Lm"o1ZeB i D5XjŔ5Tl8c\3ܴlˎ|BT5+)%׆}.!b;p=aOО#ZO`p֡ƺH˄;r߳PLR|_nT!Q37mOnQ #vT{LQTE|+jg`"6Dh;y1 JҵԆ̽"4Th%E:\Yv?EK캔kbK( ZY-7L&a,+1"a~(b &js@3n E3߻67GS}{i3cc]iILWn؏YBǵҡJ0H J1k ג8Y k!S!1<mIQMvKRMFaB$J/RhyVL\݀U$AtlZbv% D{>],E.yaJbzj5ؑ1+*l˄ɕ01 6{ Xm(HMdϗeu!gvd]dspsz曰K5m5iQM[jjZM{yaEBe7)y7.a{唬νŨ}tLk-f4&;*w,.CӘL x/!X VB^tO?Ë AI9t^P_:gaM_a#8W`y4ƍlW4fR0A7SlyJR1N. c5|B1lugWx,vksJ'[Xn#/Y[snלۻUCt@?U\<Ak哸iVhGFs4jEeVJu:EµM f@Uk )! ˇOm֑B6Ta߫JŽb;XyscBa%V˙>YhX>8A3EB$TWΎumtݰnz~uWw9CvAQD˄1Vcׯ'rqO1ɿ,XS Gc:[qn` ntbGѻ[`}35<ϓ#5U8{Wh 98 ƀqȻw%KY0D2vjKVMU=SFNNR --9 ۅR6!\|RNbo|zVn4RKF H| @)\喝qUWcNCR; ]gl#zU5C#cLԨU3fma%b3B8^-vT8iG֕V<+I I "lnk+بq_MZY0wNb}tM}t3nlGݛ߿vw/S_1֪x"-=RI ־-I W>N IGaH! &\y"X 5[֣LMq+cdc構x ytCI@yz;FQpX3׃K?(E eAhPd&5lrM(|%|B\&kd~kA؜FaҒiZN݉7Gcs|U}G1?bB}H5/7ƃL?dlH?4nz@FlR Xb+f*kJW` JWnk9TjZJW]~M?@f7ؙ7m(binr'9=TVr)914kܤp"S!!=:cӦtԬ"Nۛ0=+Ѷ΀%ߌbwikyOo,fI +xsΨE6& ?Ef"IApN%{{>3&[ũ7&lվVj0ljœ+Z EL,٘$>@dE{ͩJﻅS~:#а=;nqj* )a}-oF텋#TUIvZkbHl#BDH[]K͕hd=ނױ0#`K,O_m]$>u.a}4* "nQ&3kLd=Є\B5 JՍm1`ZcuHQ~ny`N=ƹeZs8[jfŮnI狞&K2nוSA)KP/5e6m{:ZrtDIx=n`t#I'/?]!y1/_B07|;!"m- ;s+-jG3iC~e4 =o4&m\Be?%O}OV8 w4Z;'|#]pRY=BD[ըa_-"yWjvGI|#F+LНB#nhao'[߆)zVK8{nΎf=8AmcsbR_:P(Ȅ*дmP6z >vC#Yf"̃F֓=q`v[8tcKD}FHgç=aρq6|5:̱>)4&,f&j݆7Yv^4poycf]D^g䍵dEg2<a:1Ho1񤠿.sh=Kve-iQ -3{'Y >#+ɑFDpc5wD*V{ ea݊oƆ&k?5󐝵f c =}pt|.΃lY=Lo޿:4ݠv[ CIv_'@P;>3Vvl|3ݠv@_ z X{..:sVb,߄9dF;8j7|c0 juR;fgyI"nX׀{qԎgB8c;)R#>̡H%8G ؜i}`ܖD7Ԡl-?ȴ+ewzd,='G/NO- ȉKR\Ci(NLV(&7 ;?ܷ)kv$(LC9GRBxzJ<2d5eXk[sCSX)('JkslVHv}B *6hAma5D'5nr.Z@VkƍvEd;US aePhG0h^RP ޒ%ޏ =}7=KzU~'(R\o9r|3 2ŕhThcT@ѪT5IsN,3c IG-qb[э% %ݷ&\ok9/e:_{0P BtgZ|M;"0! Ȏ`4s` ]9+R0lb&y]ގodyKpqޣ'%a<>*5s1 O&RiUq&Umsrչ30dcȆ@Fjl:s%jdɮD/H2j1YIю'#y&uL\> 4CI*alo=L1ki > 57bN,mo] 6Fw+b|,{f,_,k}a-ӏ6V!24#j\_KnKL$:q4>"'M:fAVco/FfrL*rs2 1} P@8ܜr`7i?(4zMNe,[@?17ŷ'p _pF^z?)#C0/ ZuPdoc{H Qu@}qFޑ}"$y*&$0s4-zsxL31?^fxfW{]QOK%+o$촠y.xZ!V{]˃ʞD^.GTZ/+ ~;{%bf/zvtbY;ۋ&(}'_|AoWY\}z+P?_^v|,_~XI;=*{Cw3:C5ߪo?Mڵ+ivkYG+cjm]kMU 6@E P;>F{D}s>no?n |q >ju(ujFf7#D_Ի* n]u Y9A/:T**FE9RM+œ*fc)T[յ*"XypAd%p b_!G7af| [}[݇z܉T86y}h刧xVлXloӮߴmww?$W8(8}dCdmĘŌccM|> gHu`]V>U:%\deߝV`9/xddQI RFP} V7|.`UT)W|&=gbA#/kB)˲Xb,rEg^OZpa;?TO;k7)?lo\s, Uź  J-h #D]ZUЬn$GO:W&Æ{7_z5SZ]~c<{`ĔƔ 4aNz 3:OmQb7޻3Ypgl!o#Ŀ!ڢZ($E)!]cSp;@$9_ $;x=$C̲8,nu&Cd{&>ûQwAP|[vLt>v|iNݿO0OX;a\|mǑ>N%O79ú^Cyq`RϹ[*hcYlR$sPOTc+"_}L,ב%^Yu T!$ʡ߱\mLxtqkf;X\ڎV\|XDPURƔ7F>cǐx}e.}g?|o"g={fo=n lO c3AicvoAIg~[IuμxّbF 7d&ܞ fЀ!t!޶[zdwH΄]3aL3q')]|9hY/-.OkaDS%IJ ]Ve6jV7hW]%źѕRql8&c}Ld@dڂ pQnsb|~I&X{ﺧ.f]_4.hTesl޽;zkm9.L|_O1<+={~a M.LsgɄ?:ED:k(E*7aNIGv^Pi|3GKgE#I%0oսZA }{_by檿ʷ.Р`(EO `4\$T*JUbe9!j\mЬ0SY'khd8h,PI"⠄*O-n8Ufg—ךtTsC-:3V<;mMfB J f%L&bU.RH 9]}YX(; Ն̶44 ŒɦlW1%dtNE/Alu%6nw߽:9 2E7 ):P!UleP1V*呓#XNzT85[ 9c`d/ꬫ ZM5x`w^}8Ǧ~t: 1z*4ȡJd0e0M3!Z6:Z )$5z#\\Rau=ZOD!P6z.Rvl||-(QUL6RE^es@,6Wٻ$W)9~6A;\r8lm=M[߷P4CYj#1`"UUOt(ʥmցԌ>JCRfCEc0i2ѴI#/ZxtՒE+AU٭zo\de]EÙ6$QUǢ`j@ƱE[A0yǁDx7owb!ӷ"P7h}k7dig/li^4+YSrHPD 4샹15p/:r~ẫV0笠i;N&2%gImQy*~bFD_TX6k,*ED L6ɗQdcpJTˏu># >P62Ik@/ Q8_"VLRWDׇƩXptV1V6ujjl-yCXZj_%7?ڍrMjnym3^-j\)6O1xHEU:Q(c%؊RGQ^ eKZ͠Zkb\t;HʽܚH^FJ/J _ۭ8vo8sst6yF/#հM. RyˢwhRw /g/e]Ww*nƏ؞]jDz%Imqw8;dkxrG?(}N5AeR nBPnI5쌞3J]d77i}׏W #uv?ώh*RPhQuW,"M(TP/U%kw#[:Jj&ۏ_3g%KCq}PlSBZ3n;WŮ/;/%-C8ǿ|DM!5VV+WU~MJYP :+R6Q-3x-VCjL J7L)eTN?^?>ouv,ΦΦ]럿87 QMowj:.&~iFџN&-Ij6T@R9MSZ\śof)#bRnhnO9 nܼ\hahO6r@^@95:}CH^RZ#%!\0b ٳ7k={]@_{z=lwh;@`|h?WC$ EZh*KvprzvtqvK$c -E$¼u£DQk h+w^-u4!MeO&(Xb,Z!db;sǃ6>ϧb OGk,&點-ӦMt[^s !g2s@f3Ka&ꢪC苐`=TΊ;A(u$ʮ]*jT'j7@X56K)f}FrwˋX_K*K6aLǀ,{C-7$V{P^:Zi{wưjy,EX>Կ/N+?jܥ0erți4WB,̂F?مM%.^ցh U5A&B** )JVa@'{%^h)YB.Wp2eW>YyGʍ,{DV'v"!VRJqMÿ'2ݼfA_22r W;ٓjrZTsE+п_Eb9y p @ R~x5t; $ `Yvs^0V&iHʪX=p`0ɍTHMJ[OTXx>^3ΗMw&+i?Cxp/uYڒkՌvjK ^ëxÌ1ԷvtY!!7s粱0{H<[2{Jv׏^]!((\RA ܦC&u=ȷC +rs!z}g2g.>,-!eXŒn}nuqʗxmqс I q>Ȣ{P)vpl6EKu [5_3(wMq9mȕ ٖo;X @ˋ$iހ9?ʩ=`1Je%NPh*0m2v?pKJ\jxAU,@:uP!%bC)gslOt5ϡذ7gz>ǚUnrwOȋ\PJ[υυIz^ a!{x>Yi{%w~S_-ǥꘖlVl\̆fa!w.~*ifhmn2nW75)3Geȣ1>G;Z*Nzː#"OQ&h~i[Q<~XKӈt]&l}B}|v>K+|/ ]LIf[Ԝj"LזGM )`rhLmo;ȳutӕnLN6 ? [6]/hOxRi]'ӐF|v+]hϮxcfW,wF9N{~ɮ|v77/o0-?纘͓{_E^yrD߾:񩘽_0:}5[] ~Oum6"N +o 9srԼ]61⃳~|3Yj]r,.M%.^ցh UoT:YV:(r@gXIR߸Ħtya$CEގ|I>-|6BIzM(і,-L-):`[''vHc-uYڒ8jUmDlC@YO>h`Jg?nO\75$7ӛY9}~Vf߮]5?8չb9L P? l~}*NϾkMgi>&Pͬ.2,׈ 帥nhVGPj,^͝__0:\Go^Iћ}} рO1W2=o{:WNu,?g&hj=o}ЮGfM߿_<:f1ewEcwa;m"^[مbF~I[I^cR;"̛c}f#c^9g7c?s5?r}[̕DGџg.$9La 1z1Bͥv5r٧^3;:H\E(f~inR1j mTN-T B؉񢺓0^P1X  MɳNn 0Ƹ Į~XT5}*T#.DMWX40KMH|'4q2힤iGb;xH>&LgK&$QMfr2\P*C{wO]2 n-^0TA_nEQzq:-ؘ(;rR Gms_`&h(mI(ޕBY4{b2u+Tu\px Fg }tyBLs2>01+&d?튉td+=odlu\+km ,Ăt\/mo=97wڼ#};}纋>w$ ]U3EFj_uR.t!=l8VK-`HҔkB2n]s9(;4hrfR_Y^].DjIRUeW^:0:9 r3. JMjchm(yĊk E#yd5\#Ӏ*X 5 405BQHpMdՊ]1%؛GN+V0+0i_}\ݢ͋A;$h3=a{{{:+2fZ)X9{ysg6S*/-]P-"TelAy+w4Y^+@3( Pʬl{p lVG%eڲd/M {2plV}0s^i12='HuD;RdRlKWrx&/,(cPuuϫz-htfwD쳵F,gD!#p v[Zq;0;b}? jJu>?Y6/GZoz#h%Ɂg#;(c1D\#$?}gU*Vbvo `̗+M,L% \jF}cHs F"5d=e#,New9'2c{5C7loo>^Y Fy}EA +ZsWo=\ohJqf "|TA^M_cMγc-vV-%.2;wDVK*V&ϗ⿽ZA0on:2%=%V6ަi rP,舉Ey*Ixय़R9X4@xNA|#~RVtU~ 7pc ]y!ΘǾ-fCGg~{7nCOPͷcK h_ʘ %e}}s~{ 5ݪ^G0OvN twK>?z먎x~6~ӘV禪r&@򦰦Ye/4O-ϲ+5r!hm^4fWRC(S)9Jl50MKEnCl-㻱7f (e%]l,0)6z Bʲ.Nj 9uQ#fSOpᙀ zgOb`_ӳ|86==wNX<$:;΢{=7C_6{W;-^-4m >-o_ɱL<ɥӧr6Twm/ }+B&Ca| Q93O}rɰFgHUR@En.J#$iBۭ{n}d O8BBoU޿ŶJ2g&%RYbhng_l>4es y<01嶊d9f"aHNDqƬ} @bJ{!Sü'[)}~[}8]z!us7ob-.ŰχssU x&5i㒅/toaT5;K?x!Ig+Wmg|Ĥ7}L70.]|tQ?$1LW2]|rf;X'*Eo5:Hk~ys*<yc!_ۖ\טz_$|/ ҫXf.v;s_9N䍛"=&-^^Q~{M=Y.S14q &lCM\S5 y5֮~IП0pp%F 2&㌟,!P-te\W͊wy%E&LtaU%D_2`Pl $X7&=DK5T9nUmJn-p'оjy{u1L9˪VC:r]]l(T5WYLȸ,MDcRQz]=z,(VycXd jB-irMUAiB]~$OA"K@<~b}O ·xߘ7Oǧ>;l7Me \9tL"ieuʀ˺\m䲩B45ۚ܊-#I(o))wX$B1ML{qRg P,)T\(e>QN6/& с Jn,ąmVG۲P馨=Y> T ( >PhT "B^5llDP4EҢܱ| `s.H"B_bM g.K95L~ uJ ;;3oJx&pT"Bpv.å'y#9N##{HK#F5{KNG*pi$ӎ8()}s&ɴKЧ+)h|Hvhi9qR%hEvUt>ѣ'BNkBTQ3%v=&v:r@( Lf Br|'v)S삕nNdٵ0gU*>̡wȪ&vɴ7Bɴ{ahɴs:b8y!H@L5٤ROƴ@Hyep(28L1 ɴ7Q N4턐i:bM3 ڔ:M0‰A d˺#t,S "#{RZ),?vpPAfRMemhfDy)C;Jηv+.n+E;fd>;lS?#mv:ħ@5U* })1m*o%P/49x\bCAGko1qĩ7CK!%] i¤{[ALv ѣBВTX@'F,+ x=!D=H#F*@'VGLBl tQ ѣAQC.!Ϛ?[K뇋)m{&8ژ$T*C VCe A; xiKBB$Co?S{ԾYzѡ3nJ(L{x&T$Dx>w$Γc fEP{:~9f/3w*{^樝v:#Z,ZEFu)T NcTN/j=7gG` .Ϫ(Q~[_M4ϝU| m_-yUv}';Z71gu!1tp65 ]gt?1[ }~=pb"̕lf<>@M^ ?Wo -Æ})#=煎9*vW,.fy)FՊ*+o:]oFW(y9 ExۢA˽E@I׍lĽDJ&EMV9p>Tg^ͩq.?JǩZP杚Iiph3X.C)O(V/[#dGkK2+ o}ƖtyZI¹P k\lױִ0 $i7?4!1 n?t h 8gB u4BC3áx?4e4QK5%(Ih=bZ֑JF*5I"QtTdDnI9gX IM3*i6];S&8ֈcp6gM`4>m0[ 9R`0x_wZ~piK;6 $T >mJaUNv֧Ex4QƓ˹x2t6[mg^ٛ } +qǚvtvs,0hcqLNb4{b+S`ojvNK8@=wy xqV8}4|]fk䅑.m6Ǿ`؃?"?Mw7 .ȗSt9EiЌ:ٞK#{Hܑz${qCO]-Dɚ=iY12@=W4 k]0wG.~ޅ}߾i¸{Wm{e6z+i o:@l3?jB>.^xzQ!! al猨,ZH @ eEEp\+wt5#MUʻfj,|n/VKSVX},>⁴L#z ?y ?fnh?^o|M&/^}fᖍ}3==$ HK`;N'jUQ9^*^I3}pș/\nڴȷ2ʃ Rq!XclŮȟ3ɕ{D?3Gjv|њDDr,7 .(f.^z=]݉ փeysqXhЇR 86ufL+:3hˍ&YOw!jO6/cJ3]Û?z}.0T 1x_`\ŭ LCt5g5 4)sXiktSk8?iFd( YE JmUxMb!G_6YmM{ΤZ*ޢ&,1e5@tۿJE['+5R&dꭾmL VdT }ZPik,!ax.R QD (Bi=K{#19 :mCa%bp b 7Z)ǚ7?2AN}wvH"s<#pښܺJ\㭸T4[4xw3W36;J$-cJ nDc.@pstum5$OpYCykjW2 P1+JVc V.YI0D;O8$G0j |f ?iO^iՀ4M޻¿eA AZmhĩ" G.cZoc1ZX4L0&J洿qQ_F_iNRGy8Pr=x^rKh˼vWA]%QxP"gI.6 !Dn[%5jmj msUQe cQEIP6:,*9h41REsgײX)ޖْ20CDK{XdmcU$ #k-iIXWo|yOy>.Sx#j82~Qx+M!fOj*}3m㬈`s% 8bN9ldsNjR%JbІ%<Zkq}djuM}3TpToF_cNheйK*ʥiVDxt9%msCk6ò3*rA&'33M ra,C΍%G>ɍIYe(!d@+bx J9*Ċynd4҆h F1s=dZ1*EhFRUF+ELfy&XcVh$61 VM4NS@0:G^j@R'Hcmnox1–lUZdPFjU#4žlCcaR)ʲ d.XC#$f˂\zZӁ[]lyy9$&T tw8Q:q/_upa _3/~ߎ{u}3۬Zu&uh\U\No:\nSϪljo+Hp.Xi"$e2[Vxٸd4%oFOW_g|i~srp2T+:0cgӘ8IYf)>~, {y5zqt8KRMWj|yJEe3Y!Q,ɥ ;ZB(~QEkᆲx!"~u؛4[t [%~}  `ЕJuOw%Ow%hw`+~@OEN+q6MY'f?m}5ձ$P2ˆ^u'Z?пf5/)*|*95kZx)J>oFF>KKO"\ʷ KW*(>XmY({qi50'|aBZ˦W+ ٹ$nS'EDjTmh#Fwoh%>Mj &S 3j;+ ěc+>(Id$1+G]P¯wB4b86=DBأE[xYg#6Y+ʲf$]=<yÁU_/a0}{O&QyQ͕FhnDVQfLnnzQALj}DՔHeLh wq9)SfXGq>:2ړ/-xz+De>Twcsw?g 5\m#˿Iv U5Ty.Tjt#RxGj@Lp"q{ D l&}( η? $MhsI{(OV >X-W)ҟ{I82 .!%`S8,={F"V*YsGRrQBwɃ=D3L"?>If'#Ҥd+ʝB>E^xrx7t-3??E^}Ӆr.s$PaƃR>x5 t@vGAq}}ץfRi񭨴X$>IXAk'ME8y˟X lL-o/"G$vM֔P59, `QQXT:EgL7d ߼0X&UG@8BF09J$1DYe$.At@]YcexC&q(g ,iuyfD *rF!f#" xd<(>jq8f 'wԂ+J(Ki% .3dv٘˶Ohl^s{`CfrO P@xJ%߉7,z1Mjx ġcuh%[_9GXl bx*0DYG,C9Yԓ{#yf薃6㜪$7t;^J8߄oik3)/A~Lyw/.0,;QW1F/y]8t(s ndt gO 3?.ڎomE_2Kzr@) s_/+JUD|s #:wdi$nL{vٞf #=h8M)B*0V& W/6OϞOKJGG@s-(v4θ3=12'vMԨpRVbUÞ]SZ6io+ Gq͠5o=oQOx{db9[ -d.%:?{&8qVpK P0I-Ky훩"8͆xG*0̹[!SջϜCv7lN*)G"FO7eKc"Ԇw(֣ h]}& er6I+JfgBÇJ rqVǫV0!rզ&˩$|ʦ挪lJp09^%֫9 4DIsџB T*VŚX&]f(-#%{% @.zTJ LH9͗y!BmxB.Y4Ӄ#n_~~EmٛԢnuJFE|k4m8|8 =HtxznFn$&*EPM~qIḥGR#ksߘG`EHN @bm#Bfl]FXJ0E+5aSs!It {[4l|Qt4k9^RiW<[h$Ue-XޙĘ=\4{z8jYon.DK7=teiAg0]-}t0}F>$%I݈j$qѝ瓄uٌi WWXl oV=zLT\վN?k?]If'݂Z)p?t0x;n[I).A5&(R iFg/`I؆RAtPo٘+%$^ʪSW=wz0Kl4+ V #>LX8w}Mrc!Wݟn wPU‚䛰Z,$byzy)dg+*  ,|Xy;q+3qKU0Y8ADoۨS!l&6]wĦi4f`FMlz'q.q>|"R'7Ai)ܹPλtq;o_l1B$"23Pђ @y}vv˽/6~HJ8߄T+vP8+Zbpk~I|Tm}6bl3Uٿy˻}:>ƗhD G! dx" XBe3ȝ̙eP$utjֈф>2v<+פ ƴv sq<%%5Aбg 0Q#m"hg g ؟ҺxUf0Dxn"Oec3DMmk?`N`ߟM͐w׀㜕'׆㜡̩հ_+[u,*ߍz?%UF98>C[gxk\Q}:a .os= :oEԢKKenU۷txT$`tz?VJW&فJ5=80 O rsYR)rvjz5ƹIM[ QYf+(q○_vQ+6Ztoް`FX0cra/T1"k76Kp2ɐk&v XFl1eGqd8:/uB.O7щ((L- 2 e3MR۾&}WUv¤6:(2g1=Ggs=n8#][o$+¾"Y\`$Flk5ZOqf$\$=v34=McݫXYI夀fRȖ e]2 2:Pu>Z 陷ҀmȎJn9v@5Dǻbg U`E[`0({j DKԚv1$&\?ak(KwWnCr?]]|.^/>9AX6 Y z~I",^~Wuy~.o[' Z/N%~ӕ'^fD9#[r m >oX%7WW }??n4z #X~2M6nm|˟`1y6^\ԭl˧[ ϚꖟJi"T͎{~Bc=< :R(6.dIy@Žͱg6@hVB?*B?*B?*WS Ubǡ;)Ubu}(=OII^\p )e?Ӛ7x- {R괎Alu~o.+d|S0]˹x|g=zzƙĉ,vy) mIrF ;ǔW ϖWdG0p Zͥ(eF`kAktYˬNy%evo#tAF7E#9j[7+7ʣk#2,qJں+iR~zK|:pT'> cߙ_.hw֠s?V.g7˓̌ 2@.~[{w)UA/.B &s< 1^X9!>zwVޝwgYFm>tVx+'W0 Ytk d+A˨ 6qT!!H֛iԀ$F(pU\*M<X@hHFDgh QtsK(*޲J*`yȼ:" ]03Bh{#s2k5ݞkfY1 =ѿ<]JX݆퇖B2(Dž?-4 `->D:kڞrϴrTub"@FeD:*-TV\~mi%I!ǓG 1I/IhRTzkd& ^!xOЈf_١,qKfJk߉%P>SN@t/Ls)YW W,)07\SVād5,83Lð>{I'b%Nv+Z//j.#V/w]dk'[Wp <%1TiJIE@6J@6C'"y겍PEY"Rx.|V(@~:3- `z5Jaf;Ek M,s }#<&4B?v׭nLy`S:c( a?2rdd2?-={{ݽ\!ZN@O_F>&.\*1B0kL"ڎkͤf,:C[T &ނW&࿟-p[|C C62 @mFC 8#o'%"Wހ5)In <5o~[ٻ/%G^̑7'ϑf)3 Yo ^*sRr\' 9>&͚GJ֜,v!Y\Яyu-JL[Tк4 xSA:έI$20J PiJSc)6ч`0M1S[g:[>VCf42\r:Ņ=S۫¼Wet9Bl˟%Y^Cr mKާa#$LJL9mP3'+PjTEi@v1d&^0&~ l f(s!@gSZ@>8KI(f'Ƞsk\ ΞmMJ`J偼,8ÕM{|ON7RLpAF9 YfԯF~'6cJ KlQs ˈrjR$!o&曽w ^me5{k+"̪Wr:΄uQ^/Yvy}"U7/ڻ|~Sw\La3뤔#;u9sn{Gnߣa5ܰm\m$i{)Ϝpו"i>G9*&5%`^*Yn|U+_{ޝ nnnh:cSB:nu.ڬsd-qYr,@TtQLb ϗ=ʑFUE`G@,cL6a"F9qӐyt}t}'Qg;ǠBEPa  l8dJ#-l&NX-Uͫ/frd)v]6U6@*[đCCCXy1)ƥV%tA:-[ne7?{N[O)שN6hlI~O42h>aG {jkzw54RgY Ѣ҅ja-&ݚP{fʐGQڔjRͱ,N$*`:n!C@y]̌ p6P-?zƂc[-+5@^xiK>,'X0㥉NDV<9.vό!xZMV3$U Nq5Q>c1Zxr@%gcP{D{DYCYy\ Li"vDVzk+miE`}folI2d`4ÉPc7}Ie7EIV}Ƕ(J쮧jƅ ga `+j9ͭ0zn js b0zD,#H*PjXɄ݀HlDDY{&54+n+7LTA'Mb{kAZ#0=r&F҉JGDF%f|͡γͷߜ{)wBSD)׌L#N٨(RX$ Q s# 4 Wv+v;~ѩ\(*RkP mhCPwE3< ;+lDc 8i4@ ތ1TF۔!TZ,&+!zd\Q|o ZiUL,(pr-XS58J"EPTMʝ~G~gBcmIHQCF"IxU ƈD& uKR"$ڪ ՐFJp+ bwj#*J7QsQQA/'̸W:b r!UߩygmrnwV;Qx+hk)N 6kO&طhVۺMܢdmj'z7aBjR=ox;Ʒ=) Aj1 c˰'TDa.z jgyAP%;}|CB!qmӷ kM 4@{D\~H߅1*v$̌ߋyhqǬx~gPJ7'Į5iCYwrDvMbR%YdM]\j"hי }2zE&8AIB7{T+(!Bޯ3Nkdn\jd rz?SI){H4ad~7 z, GxCPHhcKqLQMM cƤ"`vM^kΊ~-0+fJv]i0Q\l|$ s?r,6@7upCA0K'mHy{|ߣ(q)T[UdJQ)f\! m7|.rJ/"'ΉUqb+>81*cv{Qv緻ڀe pÓv0OO{O[Jx!"قԂ\ʳKX=o>qas30XQN(;"p=իoLe}*]ݼv ^]ΆA`WDMe~GhL$.9ϳ0wu:/GZ)K] ?CR˰&wӚcٞ gY}ݡkm:wu/7ӏk/tXO⢼c ܓѠd'%@_p "`J1jϕ&aNK;&tKpVyfG_!ؐ6g#bS߱"VkkDJП:{֋Wx=mx}gE4 _ 0?kx=,D!n-2U`Xj&P[U&aBhx*.Nyf\G_9z^bwĂSD'_e1+)vH6o>&Q7~-_azOO&Ҋshm h}9Xґ QuSbn5[o]$LFYSp [e.ΐAOx#j`,_ 6!Yo蚪&V P%l0&a($QTk k*Y$aIl Y$Ơ8RQpHcDs# "<_iXQEKF*{#P aɆ5 ߭ZT,BG0Ur 5,iغǬ3 0N\%Li 1b&b5fFZ4(=Ӟ6G\Wqd=VWū*[KuCP,MWǣHj &nNżՖQϝWB G_u~DE)mM lLٴ @ Fњ @D! d,@Q U`n{2\wJ pR %Sd{+ k?@lo8sH]pȆC6!9Tz Y!!AB "C2E{1HB1͙Ge &-RMT7CٰMX$ xذȆE6,R|^C$e+m++ &iIRoɊ\yf\G_j`4+mJVTF)YpɆKn%Ev%.YdEYum2‘ĖX` ILT?1.1Wᒞer)qI}Ʊ&M\CN-4qE6,ҧ@u"ٰHgٺJcF4СhıӢSIC!}ڹRp%8[MR|0.[+9[mq,aD$Bt1\I…d:!, #CCbZJ$ n$G* cK2B-Dk\qJKcTFMiڥ1*+\i4*6"Ws;(#F Agn%w@ }L, l5y.V`潄IV:ಫsv~:d2Id(6ӟb Z/J઼0w11$sW*G͆ 2BDIqbJT\du+%u@%wgƹCo05}|6VٶZ`Vb0%-jg) .ge(g+p.iZNJmClZ`g(*>IؼSRK6#iY O G*p ],rク0^ŊR3uqQ]^*cb\EKO|7ޟs/OoUKtd嶥;75AxBRÇ$\cf'?hAk3A &_:χJBP5ݝP7Uːd :O{x3W !&rX5 r;N 薍s{Zh hP(|0.LxC"^ b\"KV4=5'd/ 8v9.58Y跃,q=B~7n֛e+^#VEswց悝CXGn.1\^7kG*}h M8M?TG? Z?;h*p8=Ԫt\@چ[w͸ѩT;bnhqާx DTkrg&޿9? nuvN,u@&vkpkS[8 AJo>`0FHg"&#[@L= sXgR4֯Ӎ`Km}qi%㼡K|8E<_5=6 G`a.pB$bB[`h1"*QH &4DR356a%babAsm0G ?]SSvGEG}imb@@"!ͅl>ƋQo \bʱr8K͆J GĖ(e7!7/<#)>^Ř[tdXm3 ;c0~Z Be^ =YCL 6" d}YMlE+\}VSK$XɎکp][oǒ+_ؓPM$k}KBɷo7IIÛ3=3eBHCNWUu4ǪWpxgl^A 4[}G=ІF*_!phU X#[-{p<{ك=Nlc]u 001M̢'fB4pr؛P͚ȣG?_n_H. 6`VG{d0Dzל.8QcΜ"ND1$Q{hۇ65IfRfl1 =Wůc4M?x裞FB#铻޵_Zti,K+]"xW6]$v~H*RRYʪTVՓr:N{$EҬAx(x@Hjނq %Zz-rR1FFү4"PӮm ⚕:2x 1aʝ T`Oi9 Q֪`qG+UZو"-TF#JkKSl( i9h }E-)&IyyUJ'P,DL i0gy{# KN\Ti(E/(F_g|*9´+6-42&7KQ*OJ*J+M(r )f"ÎR-ׄEܼ&2 G-RƽSsƅ@OP lWN4 3}OdU-Y(ɨJ,b8*@Ӭ/̚h&5R=͚h=kXRV-J]\E~H=0")7fN#~hlV?yWFy4x xEєuou}h#R- F/-Q?-ˣ:GQF?ssy~f7L XےDDU_-Yx+%P >0@;G70jOǖHjMu9]AƩ f~hGx,ܐy۞H ?ͯM`6[/-}Pcb8BNIR}<<;j=)^})^}QWk=)!8D:k^H0s \#ASgtF#T(,ֳEE/_o3/Ohc^kAevKR9¢BP!ї?ߍfţ.Z| C?CK c QgRZw s]'+kx3YC֋ꬹ57tҪs/ 4WZ( R8%T n#bMX1Lnpe?S\KUL>'gvpׄJ)G30k5GGJiGq>ϠM[]53_ÎPs98#$qQr=BLj(" eY @[!tE2%SvO=A{Rw4ͭy4DPkl4Ѩ> _dίS"IU-N-bdSjjGӏA'F{%C^ȁ$痈Wp3>^n/31x(3ZA㛣T@AQJd2%qe9Q(Xɭ(VW5s^㐆hd#x;fp= 5e${û,; V@ͽ}hD{K#{heв |ùC#jhD>G gqqrx#s2hiev^3㡚L~)Qg 4A@GIeT?pCe)[,bc&zns:{dP A\XhJY0Ќ9E|5%X!Y@9v_8*swL8AF{(c2`/ؐnwC 逕R 27Xɢ}Lp Їb},b/De4I̭b sB'$1POϨ@UH.gZ ֧"=8ǧ&qG4KJ]z}uDIeB#TKyʛpѕvVc+TRFonV)k x!ю"= Wx(Xxs' * I\REswrbOh %K8R5WKOzwGWm2,_'Mi<fa.RHxS,gYM@AaȦ܍(ļ\WgZՊnqrutx҆q1*=PWoKG5-hmIB>(#Ԛ,:ϐf'bTϡkӋ^$"9uǴ^o} DRB%g+fHz0 y(hlv~鋊jK_Zy:d Q_{r9%ʥo>,c6D(A2Z !EV wa73ޣRT¢2 (F׊ÒX'!Z(#6ސ ;[Ϝ`S_(A *D \[ 㞥nRU,'2ͩCZc'!%d Vu]lj|\aۋk<.LqjÂpҬ*`N:RGϬzނn:KRvb&n!\ &`PC0Modf^ܮ~EW\P`) P *RqGAt)=H Qܮ"hkh(]9mbR{ bc!něc mT3I JpQg:b/@nܹI1RUTT\bN,@j&\sى8 ; /`K%/͠ yN)PK*|` V^(ȼt)zr$o&k}w ,%(j I/+MU8e>]0!/$r+ABJ(`(ˍQz y?Hh&eq;"Newz X!/-Ch>?BsAJ=̕\XC=3AK?( 4@3thJu"FYwUl7b$U[bCEHOG̻K}_a뷘^D؁Q&zʑǖs_Z@(̌T!%)r "ZqK3˹*ip59aF2`8vaSء rZH~A8]y-}Jr)  >?ofHu) F[7IքMv7xԊIP{Vpp-aty_Ğ~U;PĞ7Z66ZW Z׷}qY~0RΏu p{Zkz=Xe̋-2f=/ӫԋpK"jvvwӃsq2:وt# :7;Y5K&Tl7뙺2 j6kw H18X߆$;2[`Jy?7ǸX: D rP5seϗ\eݿeW-/@ Hr0Y!iX-MCHx+EkL[g8l\\5Ðooj~D6qn31Zts<&땘Z@؟^~PzRNsXuHyzvnA}.Ofi_6Y`6U* ]ݹPv"W\gdUƺnz`)#;o҈L0@K՚4pӹTui޽}AFJXjqvCE&)%/`DtTޤtVS*qJ7$la d^% %IΎf<&1ECnZe3O}D8c8!6ˤsc'ML >gvue 1*oN<%%{v62q\7GU_tk@=^47>e?1pr1jѨ,]W(cQdEv=?M٧, ch h00B\:twpT:ʶBMGW?Yvk=RG02Z&iOs]sj缮a{flEzhmNd]y(Yf=3u/>?mZS3k~~[ < Jy l2CxnhÝ7$okZmVЌ|)p@N[`Uw elǘ v3bnέ rK`ul{4:GCK|ZyhpnK<[-0x(57bpj |}5O訯F#"8N'Wۗ1lֽ;f8PH N!]2>݅x+&!KFx‰ؼ!yI̶B{2UX1@#x~A I%] 4y(D%T!gȱz /v-!|*Dϋϲ[0#;{M@zf=Y}j}UW4?^6˹z6M}^$zt)K{H*菋LjT^ig*Up)s&@Xt2 {K(R`+5 <CCz9fdk'IuQ +2,+j.L]KH(,ܣ c ! c2yba!OR(SHۜ~z*N4 zO }LZS1UOk fi=~h!N>?j V\T@e%E%!J oJq MAcU҉(qJAHB^jmTSnoE @d)SVy)/<_Ř~5?y uK!')@!/GM@!;FDv"ʚ<C!mݱ ܅ ֺ6_Mꥲ+iͶYF1uХ&@O,F;F ((,+<<.ف/B#@V6Z(xa ه@I1PSiOwcy2:!ZE_yS3R#׺"_ҺFrSԉXıgZx}vClN4dUIho?ˮ&.7lYno/װVeAˡx_ 2b-^w,58N4O//A"] 6O?Փ|ʛ6]cPu^?nZ!Z{oN9:a5 X&ŀ%2Ul@r[䘫̔}v1W5Q=ΗػHͥ'hwPN$`"/W5c._a.U3wFJ׶%mӻjls+)^ֆF撲s2u[cümئBZZegYs{!Ň_"ErmggnAy& /^IL76,]?8ۻv~,^Mczg`S,T^U1xr) ՘|ѝ{VOXV?f+M,K6jRm&k<JIE 7*.xw9xwd`5@MwAq(gaK^,Θ %F˅IMZsQECj=da4~z#7κq+7Qݨʱg]xy.:[.*agV־UOٯjOnWL4Z3QE3gAH}:MᆒR滊/MnF|1_`%0`*YQ & ;`!쌣 сVVڊ{^I6%[s&SMT)~=FN)foZP%iQFas YY@a,7WJdSzsc.4vz^Ŵ YEϝM&/lLT~ 76:[>Rу%q:a5S/muҪU6X_qzugq%{Mܬv6Yvuo^5R %cťj_W|-hy3LVivV m9UdtbG.&dnNy?Mb@xBjGӸ9%2dT EdPSrrǨB\IbfG6%dc7*0">F8):5Z|p0$*A41Jb3guRaY}D.Cظ{D`G}~WHtd2>) Aל@yRHV(A@*𼩌6b#/d#3rz VZA3cŁRv .I v87w54Z x ͎[;6֧熖A-D |1pbWya z45b+cلueջ(waU>֐+Bt)_ 9RC̔-Sj-YN$wM+J ;wr;$8',ÕOR\ w &Q{u$Xk/x> O ^܃nqCHG`jh D!, Gݑ?6$>,Gf2ֶ}ࡎ$0&LNFQ cMpGx%h oo(!9P"sVaĽ<̛|1, msY'eȊ5o1@&xP9J*5b cH._ K ۩{>a7vL_ٝTc 0K:dSFuS _.vP&,m$ H$U(w(9 kceec89FҨnؿc"HA5 lh+.2D%CK4J5! XfjH0F>HZYy"QHզX!Q h:sGb31_a$V73yasL$i>3W/*a=u'v4}Y{j1Y=N)'_}&C,,Kf"Hͷ,9k_9))·+)sY-+^cڃ[T{ y #п;~]{&W]o?n8^-]ߏGI1}\ݾfW.}ϓ7wuK<\Rl5C]>osGLoF7ٳ@.e->K]=f ͐[jgLl-D~ZK8~3C1IeQp/n~盫 }Ci9)>a͢S8]3fSMwF׽z-*&f_H7z?Ɍ1[Aю-l@5I5|w%o2_PGhڑb8hLiJmRDk`Qhq8] ?xdo4PFNJnU{GzB1gJHd#WaM'w%(wI1shMU7t3ԇ1Ήϣߺ^ ǯS3uB9^ |W, e =Ʃ'WiҺ[Y,/V[BfB}EsVMk1|}׺_y^4!]S&vnze aCUK3 }oԽ_;5u!.K#ѫ7?M'oI;kפֿ;=H{7tlGnlz#)v5_fl'kA7(_^hÍ$B|[>ؾUE@hKP۠d)jAcJqF5}繂z;x.ܮ‡>V&~Nn}܏0Hյ !tm0IOPJCr x6ܟSYmu\n|rx2zp-mA2V3pc='iFuZN74h(!EnRC +Xʶl`1&DI+˙^[j ) `ybNAu?!a4J ADdPDR׹VuZ;kV{Su>!2xQ!Ѧ@0>jvʘG$~iէE0![=3>#6\/A-5;E=zIMs0G'Z%%֙m)g<>P:lo]Bѭ}돊0\ ϕVwnbIh H`+m#In#JyzclOz`iqZ"$aoTJ*)MeSYeƕ6LUDe*dM35e(đ&Rs3$Pj0WG5Gݕ1,W}'ҬNseQfq0ϸ9VgC^hђbHvѵY!"!̶Av\#N7>6>>c| G5['*$e\5%JJ.w t}H](P&PJ  JC\iӋŠjث4 4;+R+$*eKp$Ǥx$Q-X)9}Um_|I4  % 75D2\pVQ%eBLoIZ3FbP&h F+nfm$!8EӶN ;<&f4- u1趮4QKSNc4`,/ ӝRN ;9刪ȉ'JUI DdS ~+C2ۼ .QT.khRߖϨ]k^qruCJ\Ȳos7pn2Ha0A{g;m\X;DzFzS|V;Fs(KίQD zӽu syZFo:R.aJLeԼcep*׻ƥ+_ʬwc,ר3:>gy+r7,/Gd=|> ؙemCRX4" Y .;YH{ufesY ~pPrKuX];k޶[-pnꭖjUcuKqВC+o~o aL[Pʅ6Pj W - XlKho+iA:39D ] &Wʻ8f93`c ~7xbєraYtY2|Z9?dX2ZN4#B {YpU2;\j}e*1%bnNXLJ}jT((;_Q);;NԌoZ)n۵)1D*Z%>  Rԇo9~Av5`T%`I!t#߽7/&b|<#Gr(ˑ,8zmVcBYɈCED+E5FȂ2^c"@yadRsvS)~:I+_:JSr[m5maӁqǵ1JZAgμΨ5:Z@{$]15FͰiSlbUm<̅UV/tIKXf%uQH /N ݵtݩw껧^[h.a'HRkZ]S8R e{v?b'+y<4uU~G>*֋ύOsGYi Q XQBnA,Bwt |JBp EH%{ O$ǒ EfW3Ycck)AypD CAj˕5E>FO"J2Μ#A g!3AjQhp D'5ȗ\43= aHpjcM<  ` A9aL.sՌE 85V`(ቢɡ4 Y,S*-J1m_6kc:C179Sdr o' \l'es qA:,2$6RƵAe,{Nh&&jŀ>r`ր@GUkԠRdmؑ5ֽHPLAd<)SccB2uz5uU4 7$E̖&@1Tb[Q9G杒RY%1(*MȘ^`lյBL>؊hᜦkc@f+$}R; H S&9R"9pRɵ FVOѼ,/@ dLQ^%k'NZJ$ 2kkJTJdjISZB'go0ΞB&gWgO!nrጴWJwmͽꡝ[Sx:F"^R&q F轣zQ"(L3d։28DByThMe($pG&$ֺ;ufbV0Tl*q$}R*?a\qXhp-1j*Lm}8(18!(ppSl'etypL+lp uQTdIi8Jy N!Fl12eUV:.d׵>j@T{Eu0` 4nB6P=l.:5FMFs'eB` sAʙcc :"ae@,-1cdےQD VN#-LGac-@ [k,(q\>,brFb^.6}yˣ0zyMC7jow/Ǟ~}]k 4\,iPy _uhؔUS `q>o[Qȷ9GWej1@AH2& $2;$) +! 8 KBp~^[`!'?XԌ]MJ:qU R^:d3Agf)r ÌD*e<y(GvpXVGk5HI>] PrJJ@te*sEM*u0*Uqɶ9K'r"\d?:GdJgU.SPE#ߦtE*xAyK+]B!no0_H5^?w3WaIΡ^?zI1_g?,Oӳr5\2&N/r?})1ד_jQ{ A3pl˽yL`W΁+{IfYepkd+bݒ.Ϛ8qw_~2zw~q:G>Lo Z av3wm9Z7(귡276rO|].|ex!K1QHi͋_-\~R0w߯i]Ls姫9Zs,hڦٙmdݩJky h;\.w筽@Iـ 8ݶ WPanN :g>Dg׆o+z2blsbFԅw2*$ӂwYc7'g`dcq Z*d.'OLb`6ygO;XtYxg ygP.%wF_>P`X! 󞺌hJPf`ݻbr[/H_#j\hR"5k'x #0h-58Fh  =DGWs^:x8rwJC5 \ynEMQ- 7 I-x!kP7FT !0NDB6/\0dpgGr&//N,{YxLP@Qڬ * 366r0I1vJν!k:̓Zyg[Fn?W0β_^ޛߵCX D-6pB,ɢ3IXr8H"3+\1c1X`J"Z{PcSh8DڨO =ɢS()7^[ 1DS}ĐLpPds9e=u OrR֧rRu !6#xmk&Za"g,l'i\#f#a1reqPbؠ\CGM"J2Μ#JW!@( B1Z86E)VZ Hj8>pԵs:8lA %&$Hh'( T౐rK5"sP8Yc M'q!g!N=cwf9v?āQd\RrcdK){L-|=.Ԝh]4aTRhd>lYlˊCE,WxϽ6ih_D^tmL6w/o/.4Ǧ'a~\M]2tٛS[d+/`Z\%3k 3Fn$Lb~u4FK.=і ް0H7.Md}VNF\R5A%ixZb}?؛iVn-H7.dT1ɞvaE/[ST.gZbϦ#ZU!!߸6){n C5A%ixZb(SrFڭ EL8,-vddڐ^,oxJ"ӦVL[o\DT)ZȞv[ST.Yߒ/Ѫڭ EtGpE#TǪL1'N/&F9Y-y->e;Sjtnf Mqۼ[|➻HӊHqyяx{Xlh;$m/Y㫯4ijæA4V~_{w,r~8YVw^$*DXFqfG!º!DHHBC)8p Ô9a핧&EbRwlMUxkA%8_̧̾bŒ)IZm`2]@̆yd9rƨ i6b3}'H ]U3/+ 'v=TM3\o*׿$6؊7|uWg~6ysbYڥ_Xf6#l mlVPhT!sZL"`XAuC[2Z`CBSXO j s8:+4bC+][Nj *CGl7\Dڹ@ [zA2|kO)إX  IX@[[ 6a ~bW~b=V֢OlOJƆg;΄cC&ZL!m&C8zJ -+~M]eE6ɬYPsiqioȏ5gS,h4vÖޅaYMrX=C.RRbsfa2yΧm= 9OQc~KIBިE2L; oyK^x:J'UZ/`X4y ߾0ʄU4@1@\|9L⹯$ Ph6:* MP"BjDBru0)l{n2yw:kzW !$쌨sŰ7?Y I6|oxٸt}}'뇁~Yٜ뫗j;Mwe|օ#Ay6州#eB*7&콱̎Lbb)C~a-cdOV6!j{&FP##M( vXf뻍!*aE+H X0F.3)^d~RGe\> $=PCPkJHiH "U֫X P|գ0jrz4dxzQqգC$za{wzt\&n* AhDO2S;c5.2 E2N$2P$8ҁ}K^iq fqc(Dtr"&p-P0hH>6vM[0R6HqM3JAt [`CHM>`=M@xHi6PDp"@6(u1jG>Z+ ZGJ)*2"&EjO;bQlD>nUFK3gjyZ:%>9M*"EFD0{wӡWJB^dQ~|w$-\a!5#{(2a/Iޗ~, (  P/߀ӒqkRl佞v9{3YY'8#!Wd36_Gl>Uk='ͭƭ5-s "3Jat5l B#LhPug0BzaRuݫ\Xtgؤ-K ߋd1֮9Lc`T*YIh:^r[}pK\OHZ숳؅ި8G:;>Of޷:wxy^BtvfK 8с?ݹBBM3s3K eJ~`5}"-{[ q a4X{>SIދDws)mZj#yNP wJիN'PgՆuJ- dny^ 5f:vI-!@3^hQ'r}>4rWva9sƁ0J[Kf3hIW֪w@g4ʦfVZ vh<ѕ7 "5jNM-'"˵(6U.BWb>jn^8lȷ G8z3?[>\dʙ%˰(r N fky ŀ=N'K鄳e _ B{ɸwkI$=}x"cu H#XK|Krtqۨ1lַaǫzo EշF)9&7;u0" G7çbq,]\NƁ@78L7upGaK?T0Ydb V$)DcpWX~<_t^˳I.Er{:RzμFǗT4D$咰13˻ywE2?.H][wڄYгyy1 \h&/Q ?PJJ: ͳ*O tqsuY]K*hoٜ?D`n287eh}MXOOٞ[q;'t2uHEXcY'VOr8@$~O:/rCnr:Й@eA_nưýezdwUn}@*N@7PAzC{͑KT*Au8BnImfarR,ޑq\ʂ_ڠ]iC>[HM ^#^aK.;$š$w93KAY;C;;݃k8 RU@5@ڍ񳷶cf{H]H(`vrj.Q Y89doJ>8a2;&:DP&)Ho1Xp_?|^%-;Y)U! L?o HhyOޥNӹ@F9mG\*&xNQ+>G!@a3v[6!vHplNd#L),MӿhhaX)ok5W25 'p*'9D`)E8n R^fP w8o$HS˅Qbi} "E4DߎimAgS|vA!qN ;PQʃ{ T,?#S_φہ Q۪R@D](O*Ue-Z6B.R7^hӎ0oSY~%g#NEn]zCff})C ΋MJSaO`KIw Fq8B;<{=o('M͋pӮs&YmN"`'Om:Cٝ)AD n^*DB4)m!M\b.پ*Eaj  .矅T7@BExX~l?$(gKcK ~h͑D\%h%3y$)_TЦiNI ֊WW>+ `bFO, kƘqD׶6<~;Y怎DmHoFd;l,P>:D$DAk72&pKI li:fR5qn"&~ʴIe6_B hGg…_ vl!ɶ9eb$9sHX{"k;6V*7\phXֻCApú֧ 6qじDꃏ!˪ r~n#<1|Dw[tTЖLTzo+p^~ĚA@BTxo!J,&kN.:l0a:}sW`Fv p\:+/[};ӲW_QV}4c 0b!wz7fCޘ g^h8!X p@FO7OgK~^ =)"QvWwƫ-ļ6$.r3i鷣߰FH ^Z\|d0:;QBzӯzlVVd6A_Lc6f ~W'L/WYofwۿZzU@ LHBw׭3wJAgRhR߿}{LǶ,Mihl4-Q$2"F4.'He3&ngӑ囥}z?CKi540Wn\*ܳĬ`0O ZWWF/\ܺ)ϟ>sYL5td XH "RP c` 2 F ;椝NjyП]#H֑ @\&+bIX-ӔT%;7HH% !<&#H ,bkEL˿L4c##'z`=_A(ήW*'b |u: $Z},DhJOFH(!\(echC\is.\*D_qoޮhOmFk{(C?>TTPeC(5Jgή}=J"G9CUoߏe cFQ뗋QvϵQHEk4;AɆ d09-~ₓ= A F* yV<ކauG.LHX".VS,pAk`?7˙%R%#`VxH- y9 HTo+e :|81Os.PAPg~_-Je+a0+w[ژWz̽R ֻ7 *eqV&ڎQ;v0*ϝ@ p Fwil>YٌFd{6\HTWuM=Fy#KI[R˖ gLHVlؤMz3vo~>7Hga4YWYӨyJH(/.5TKk+843ѣ,NboN 'HI~w[U[ṛ ձR?:FY)#X"HELLbQ(V(#H!K {$6`QκTm/!VO`rXi/#6ٌvʳztܳujQw̢ڴ4VJ5;05gfǿ}l݀t#*~\u*Qm۔l)q*Q+%t9q2~s?uҋݎ,UjWJ6=󧠜0湯`[Ƀn_,yKNM(46a8YwB,TIKk'ۉqVJT<]g-ݳƟQ%hkF"y>@/gnI4_DZ+t'S5 6c)UgyI.J`55ZOԫhOlmpEğ'Nu/xC*n ;K΃djJz*Yډ=zHYۙЬp޵\$ Z>uRIr~g(0]sw 1lO4 cpM ^cكq Yʒq 8Sv0Nm9XM=f@wwp);sZY=%gJw-%lߨ?;vI$9tw`,ĈAj:vUP 89Cܣt81^tc^(*ym7"h&D uQP4Ylϑ7$_ܥ7օ2Ȗāh-O" o6-o,)fEm?M"M/(AZD>>IWIuo?Ҍ35gl C!On_r8QM7BQb]DﵜGb.R(GN)(.L-% 46C $A4P0#+ZWEjtfCp/1lj5WKn_k %W3sgb| "bXe~e9/!:5Ea>Y<8jTl<5fj7B0@ys^UOt4uA k&mz$lo坶3ƋiM!0c2AkFjr6d?Xf|2rsѦEw_'Jg4 ^jΙ0{?˵vc<VHqs,2gT9-D.2"%@ JLJRO-6H %}w."*eb)2{5\>fd5#/Ee2/׌f-?⧳:X6f9{$Dk6K}P\᲻HԂ9tCmɉJ %Pat`,E nJT t_L°ZiMS[-H?[at}1v:=pc7&=pc7vz[T%ڈPù"0GJ'\f"!Iy2)S^0ׅOgkJuɑϚ0`0r Bs :pc7&pc7vF,RPBRRBCMsdsɯk-摦d1S `J/r5}we}5$z+HzO?hC؍&ӲL q2ڊl6S2sdQzb+5sS?G~ʧ^;-74uE l1XѧQ֤+)s-%.[/.[#5RP)EBA%C]+/qH`*p*QyHqe1 <) ͠@U@,NJgJNpJOg1SD8ѬIX>3yʒl< &Cտ]= 8GȲNZt0^-d(XYeWW$qȷCt>ugu VNuL|]] *f12վMpwCÌhECgU!&Dٜ Zj@ N.s{𛊵PZ2 3Ҋva?`6 @`>NPZ*ź>n]*RDW7:hfOa>T);xCmZ_I:ulx|MPviTQJGG)DZSٳi3`'\G4|r8V2C Q-эv;wf.\Gq<Sn2Yxd_dX4}3.5,@KpC8atvR=Sg_0YWAGr"ڷd$cJn/;^w_`/ Rd%ˮ0\w~V[FRo;Fw{AM?xm%*:P댆A!jNCչؒk5`?,1+L17}emEK)^P}t6!,|A9o8c/]R̮xD/,rRzi6kH 䵝OI0yrأudRԘEn FgKK@;ܺm_ѫ~T0(I$"!Rʞ_of#~C,ROk* I]űN]wCe" aYN}*-[۷^)a3g'm OcI;jS5`EY~jP2G^3g梵9d}#ۅ-*U29_%ۘ?H C uWT rK 寿^iH't{|&1^?/ʿ>7"* z=^e>6L3#В֏5㚴l+j"%%P%*-n&D9Ӝ9$uJ`BM"8gi( (ፄ3d=[#js~@}H}qr{ΊR?$ E@sh#0ۛu)j;Ow|}kt}5#Ȧ]}`!J@G\wyʋ b#֕[ØcbV]- iqX 30@,ks oX/r 8nxڽMwH$Ķ#ύ#{"J9Ǐ~BvdmE\W#.qa+H#HL.1;K(??Y! כOa-_͝kVl~C˄#tM v5e6;|eWW7(>a~w3+#*ە_߀gQګ|zij7{HbI>VƤ+C{]7767K꼡7g᛾'s\5bNtvHSog\湅W1r +y[BJl2tnmvGLχhaon&ګgVgR%2rϰX>ܺ^aM8y^"f?IF (!>1…jxyy~^T~t쒑Z7.0(eqxha|wyU6OYqR\R}3(5Җ:gpOR&%pOR. IIܧ2ww?^8~eEƾsƋUyūmۡ_fw,znkG׳̹h4ku&'!WDjS#*L {Ge(cƇ"PF x"@`4jY gf4~z;Ԁ3?E]%' 35ovԒ~1]LҰk4idd^Zq]pAޘmO[{Mz )`>ɕE迨qCpI]Oqs`D TX*W'H +9acTprN'Pt1ӎ.XATHa?2e(&'qصvQd\a10̹[!.4"$q`0cJJD @PmOL'&?9bB8UaM]oymRZhLqV_(&bKcPDaLI 2`98Pb.c)iٞT7bR[q%u9CA#$& )(R<sqP:X\zIMi+NI:n`a w J U%N(Bl1Z FjꜭtYRIIv8&\%*dxʀn+ڋh|i"%"5 vR`n%6`HMVZFIH'$<6THvjO2Me"rn8hhBkG Fpd(^^ziD6.%qCrTZA} HÈ:[<GwirCd+b&tOhwN!F+&[q+mֆލ%IT*\H֜ci7ːZ}%;'K{Uoi 2C4GVLHG+n. +bZCqCq?+IѢʼnw Y҈׺U??nܑ Njv @xt{}퇘r>W݆QS7n )5r`(o?{g7@wy,Ǧ&Ya?.Ee57bLjįQ_=+miVO U7"VFgqlB6|RGN s6s0a0IV~ ٻr4$bO >sOFNMXg0k,8Nogsw/euJ8z+ռ&dcX \9U2'-$d|E!K9c|P $EȰ_a"8xb3NK^xq(r7VRR[*fDI ,]n$G֫{׾`kQ!R=rbS:սkz0JJ']G DTy"s30oRn4AHzJȔ,#v$AcJ W-2ex0T$}ѰRShXAvPk{@)$J!by(BY_- |Lj VqS2`Hy7=wZ&:m-a8dɒ9,)=ce"x-AkO]4ĐB/B$8#-#唗J:E I-eLHh4uXaD; [KKkI qR ^RRBd%_0+9i9q+F4(jIS;|)ϨӖL 0jD!YTjC 4S]!sL2K<䀔\dNvi3ɬRl/Ikb{C`-EoG mIG`Oc$+?I\K'y8+頲2@KD$ !szN{ ZK#q$i8RuH0/ٖKI)D1Ojfs+mʆ[I-]yTr89^;?Ki]7KZ<G^n^;x'DeDeF׮4:B1,&iVKmQmd`*íM1D] 7 Fq,S,V].1P$rU!!@?B # E%ҩ&L\wOjugB1L hX{ b{}_߇:Anj]'zuz0?ɁDBc2H#9ZVq C {H)LpL@ͫs̆-7]N:sFp5CA B`DDPmkVS=cK[ynED^*uHi!sșvdؑ1FQt9&p(fJbqHdUH{t9>rE[ bGof YF#7 G۶i@{?v}oדqyyun%ӻ޼;y%0kdCBjuٖs!G`MX a`Y:acTprNf@M̗nbJk {!xuRZF3%4% eЁK0>3n޳q$W}98q|)ikJ&!)S4Ù!)[C"^]0Vv'O7 [}O iBcsv:Z{;4"7ېyUֲ],ihf%d^1  H S[ ;nY |y 5ҍ pooq8 v3\{$cFXH KH(h0٬\Bɢ\dW^[Ԏ {Xjk͕k|͏MyLnA\@Ќ%Pq0D9FH =/"gƥoqk*j Ԥ-Sf`K-q]m*u?hj|=OQb'^G£AC ]Vf_,K6Vhs!E*;3 vRZqilYhB#,*ٙ B* dC Ėeso&4Хܖ 즞5Ed[ԇ )-p*DEc'"*Iu&J4)%?ky [jDiaߑX""&6MYt5C(pǒL/+ !WD̕P;MC|&тG'~5&eM~{oŠۻU!UFq|zk߻Ou?43w* jC]rr|y5"~>99>xxDWgNgW?s~KşoM ?=9=#Eȃ~?k-V-iԚl"NЯ ~njg,Y!R=@dC`CQ!:-!'DͩB CAe#z a,/:C]wc.{kY%Fږ@9pwAǺ ڬ hP |ǐo84ډE C Ͽo71۬0ء9߮Q[0:=0<| E;䵘1M+:MC?G*v1H.CjU<*^\3&b)Ԝ+7"{]yzRG_ zPFsMQ|zVGMe >_]<;][,эFXO%d(`C%dɖQX%<>l1v;)[FaK%tdǒFLEl ɠ,ؓ!Ywr,QzoaLX>N]~?;nY]뻖S(| 'ȏ1.xU?]-)WNa}W2mVm^Luݪ[)C6ӊ-kI?_4ΜVSpU+yfJʧJQu4 W:pH\y,OS_=yצ1#8a{jY5(/߆;l:REpZ/xCn T+2NmXY'#|FNWn8(ki$ӳ'=P@GGggYXOY6:r 8W Jw-7xÓ]1bwFawFyǷ.{mŢAsf k[qaYt=*_|'gRn8- Em\SIzA3)a{/}BBf$ޘIxE"AQ>vZ/&p4Q@ TB{ğ&o//qE&s' =e!ocA❓xOK3r80Q:䨈R̶^)J@&V" X{(* 浭0t"7zil=dM?{ѐYA^CF+7q9sIFHEȄ2YyoLR[XwFʬ 4tultŸ . ܂CqD6.c#tFWτ@۸s/1 YĄf 8!R5(ʐse.-~:iϡ sgd7ݝK:c6Zo]R{Tj5AE~_깿[`}Ց7$Sw]wzE/uElGy8'HVwIst C~Cz`-Gh}Cy聁ѣԍ 3$^'uԱ"LC"7eXYS\4Ej#i{@lt֥w39"hS:XTdS;L~VlZ}ed̄I|+ܻJoKyuㆁ-r^4U'({P=j(eݗ4wա0rE#uNÔB{H,elQJٻQM.QM.5Ym2$B-!F|*&+R,?<)EԾPZ}Qپ/XI46diYVՁ/Z! C=:J%x>Y6dV">R:ۉHek7Rb=U:ۛ!gIe}b2Uw $zw XG`KphvֻQH@^ ʡ*G+6"q=#!Pb ׏|'(UFojك6ӒپZhV"O(&:(JQ^RN0EHUs>_KQ+=%EKHH‡ @ɳբc*JY H짠ј xצ P*)Q%͌V9[Ҳ8ex6 :HN|RB *1T)"#"®Bz1U$F,m!G'D1hmBU "E։晆K]}GF 1ՄledY"47g h1CuJ`jhZ%&nX-0 [/hAe5A~8q4J*}H4C>Q:kf'EUr⊟H~{oZ ~jA{wԡh_>~޾k^-Vx}r| VBT|˫-C~>99>ڈ\?'?c k f_zY@M,/ggx->;*Z#R^iPJQ"}&Pr޵;nj)o#9UlxQYccD SϬ0xV2H2To2Qng!VoB}!F -mu* 3񳐆V[إ;X.Y""Z:[[156@-)V5$յтif hvyMrN&5Av=Uj j3~KjAȿ$ggȿӬ_表DB´<' )!@%4˒G-%f8+bXFQ34ǪYYpYfC${\E`HL.¼l5P A}Eŋr0Y[ s8?<F k *Oٟ?} tZߗ;/,HnAtx 1W_{d^KO/3?ϜU ?Z: 8W<9xB l":6%(lmGv5 -:Wt[WXU)С\WQvqR=+ȇ>؅r%W\v8_ZNZ`O%Qb8rЩ&jT'GV4]^SAZvѡ*P3-l~j ղ ɴ抭oN8eh;U'8b1 JiEd4홁we=nHR(/;i,0itlddj)WcSR)JL!aWH2#b0"x)Jt^+j/M? nT4T_c4M>e3/]D`I0Gc!M~8[Iۭ#AfRSc/i:zv~F>՝lErd'M*&,-/LtRcT3`.rq^$y!5}CeBUȧ"FQ8J%u4FG\&<¥^*l[jˎԔ/؜`D0vv $ 1hiMQ4(!p»!k2p0$v! DXgYLӘX̧RGQbS*"SHHJG# NQXQTN1%dsDDb#gG&mC9{(OWNC_RS*=%?<Lppˏc2Tە!5fP~J2JW?/injKaYVs?wۣf/G-Gkq ;+ Cb{"|S)˹?;D$>ulbs1Of~e{B0wddǃ,v?9@~>+4KQ4Kbc r%hg8@8 dp! xlO XP i$6H&V2Gi+൚?鴍lD1ٷܹ^Yy{8)jOsЉt 0x>˙ i qqv_+`5s; ぢq0hy>ݛ뽒U67@w _G?^}929>{H4Zrqo {϶Ż3t#5DCߖ[dY2OjSWb)SBol:\OKą $ ʖǤ{io˳ToxP~w}@"-+;eݶFE3yo:hVRFn&[G_GtⲳL{l{&oYyQmJDǨ[o~v]mŠCs8~ ǎP"4!cT8a3c)dDzF+WK6^֘=Dz7+֌֭@℘˥?|1hfKE"D{ݒFiidB@&1s c<7 gl5pyTu؏:2al et է&v>=h&(W-H[z]x<.m&ɰ-Moq4o˙k6FA j/\]&0+ :3ː5TMٸH00"Vsk(Íi%3iJfتԐإ*iT%W5dm-}M|M1m+p ΤHRG1f&Tڄ׈"[Sne39}+,.ǪzzZ=؋*Zr&W9D L+&I\MLa L{\g+J),MJ4yS=-&r }vlǂ+e!;< L|FhpV&%WAӁp:EJkb:l|Gr*1 h xD ‘6;i"ZE9Eg%`SJCh+LޣoPb+>r4N|.[71k16goɜ  }^\Mÿo^Wݟu/֝_#Cgn>2[b2>5>̷}—z?(wp@`r4Zdq6{D>~E[_` #&~=P|ޞ#c(z_.s>~?YwdEϐЋ en0l&^vQD VaͫȒ4qI uv 柃Vh5W+x/Lܓ3\G{EO #}7y7 hX*EX4Ŕ[2Ji$ܚDED`w*[Me*MIyj~ͱxzXm!$)<;׼oπcEdz- y=FIK#C-MwAE jr\?.*{-2hx.q"Ff!{X-|hf3[dȚS:ĜQnLuBґSwIP%41JIȀ>#$IژiGt(FRCɞbø"?;̻7-{5@2mNU bn!, D W.zUfRiN^ fB2S*+A$L0M1liB_Q0*Ei&gg ņe#CiaI$xroO_dYW5\*SS@qmye[ 4e`4*HCj.ZvBܧ :S6tnW/d]a֊0R'v%LVﭔ2`R(s_X#vRB[Fɋ/1ೇvޘՂuD.\^e2-`O~ #c.n^tg˝U7J{z9Tukr n^hKgū$Ib>iY׈{iN`$vSov&/&ab"pbX"O\Q%:A)O|7"deH8}IjX"D'L0j,e NaHj^>D`SnƇq3rh[f*p~bG c-Qb0b{S;˹}Onx0/d'ا̞S'x̲=jyGgwddǃ,v?g<M^i +(%_a(4ܲ`/[%2ysa=̲ɟ]"5N`%!ICX #DSl|Eҫ7w/n4$uM -%W7&@gb-g Clb!O$b[(KŎ$MRRU-X'@SBoiJU"f2D,Į & D&֒b8.D`$8 TCVMbCcqYB2= 0l>i9"L$#HuXeBQ6z0I5(vOcd% U"V텈l=m=E`%r*¤"G4R^[+hr ˢY~Bd6tg"!C@t Pj!04nՑ 赾Q(z?{UuˁRpu/m"J|U#Ej կϵBTV?(Q\N/Gj 'Uսn@庺 2u!ԭkzE'79߻ףANQrpo|6GKG|&OuHWBr3CKtso: R:%VZs2IZ:Qr/"Eܘ rZXhPA7"8 Y!8%}[=#0*X}=#;ꞌ( ߏc0F]TĞ8ijxs3^w' &ϧvy+ ࣣ.ësιՍ(1g]c 1Lc[*r.Ո[kc,&6u*6f̐O٨G qT[SIn.&낱;Gƍ_P{,>PWG;6 0kZo*cj/T` cیPa3v#UGVȖ) @ }OmLW84c-џ93[v)s@!cմO}ႫK΢< b.c2)Tmɽ|/4]Ye`pA^zF7JdѦoeׅᷣL6x߇!Ęamᶓ09~-~۝T>j {9몀0`r;{=A4c_ib+@c4vG@gz8_!lFŀlo  9WkYIv俟J=FilJtUMkr5Ӛi|vYElԈ `mNNbvիs^񓰚 V?9'˳Иly6Γ5ARy54H U$09Ȝ>%!m扈4.H"A4鿰M5"%0].H.Ѣ (u Rw]uqi]eZw64+Hկ-ۈBJQ:6Rl9g)F$q 1JF( R]'TA䇯mT洹o~ݩm?O~z}"T'ꃟ~uۛ8k48ou8~ZFe>KgJ<(AF U[aD uDcc%D$-J d]jJ5$P;nMPA')H }噹\TgŕNY?ߴYmY")0gdmѲồ˓y0㛏Z`yx7[,JSQrIGѺq#糇ßZӚB6_7AO[}!  jWϧ:5vܕ2>`gJaʭD;v(+;p6JL=[,'@X7b@RGHEi!;y\S4}Rw{` Nz*}&|n#䵦){?s#s;d$qHfK=Mߒ>7G, go"|`gǛ@Omcq[|.JMnd$ϋ<- ǵo7?9ZkY.W^,;L2I,؊ʐxzb 9+r8AB+<$?Ĥ DUD0ۭx2%0Ty r+GyљAp9p24OA dQrvKA.\eH7*3kU{esm0r']êXe](F1!D:88x/f9nY!%P! aXFLgi$hĊK#uFS-#<~pJęu&QVw+:~)IYz΂H/79ǩSZj4H;l8(eP&%L-p"~T(Р)dZsd Ҵ6H Mk4mfr9"̃m0{jSRVHA %gcC@]'\uWZb2rog$L9ɉ QYK1"đZ,F00l , FP YrD(d1$F1R("q`<ԯ԰ yN TNxYܣaסLS캱$PS%o>gV6e7|^+- ?~Q󞜄/+!/QngYU[ $TE"$!\ᵆ|qƒb}-Xm]"l]= 0bCIcp^rp3 Pz"jq v0CD)xLWW:]1]MorvQ~0~Y}p}s(a@on̗:g\o~C޼ _=/V#շ~}0ᇫ'Q!ZWaQ!3^_A]O X)=_`NUQ$,&DK99P )>\B ~]ÑS !4R BUpKژ1 f$'@ h XykAkӻ_B6rs`#RHçz~{*ķNLTSpUV".vE{Sl?̲9rAv.H"9)d@J:Jb\% >^4˳vM!ZGnXhKrׯoXz`oZ}ZM Wa=PR0-X"_;yکh BΗWBt9c#[! ˬJ؎k{>}׋)<,z/X-I-YiiLD02'rӾD6]Z 9yAscs=IӵbC0%مgu]D2&c]-$TC~W 4vܷ7q>? <KeC~sg:==vi#jZ>2Ⱥv.0)c֨A~>|g]Z&䆸hbh*bU[zt;g*nx?)sK`fBĠnAZ'4cU qy3n$ϐ>6`raJpycqCz뾏zy=+=s=;&)&2V>ڵѐc??ZsׁoJO?^K"9sq_wJXygoœ,Qhvb>?Lg!ZC4)X.W]Tg?ڂ:MF,I厱VR-<]b}:_:[EDlwJn%oM@gnV\ݘhX%>&Q79)3\-NQJDc蝮dݓЍBu倨O4/OT#G. ZxĀeTvHZƱ*z+&¬Š!.>Cra7ڥVYfe}G^9D},l$urggzQW mZn!po~.%NqlH\7]dwrܞ3NrS.Fӳ.>83x)7N}R\'ʽ7v'i!oG%'ζ A]#wl+yNeJ.@%V'kLBP&*H"*e*e f1l,)s Yz;25UM-e'c1r0Sk5IDFEҰ1V 3P2( dzЈ{V7 gxkHԣ]uhvC#Ԡ̽rko7*Av6o݃>Ҙ VhuN>s5Ybh5ט>/Wz`QrCűCB鵫"HcxG&/c1D9Fz61(rp e`~Q<гx,P48ΰ3f5 >xCP (8)[xm${⾷%EaHJ"YqB^JY*^lŋ݊r~tRG&eb^bPܑ–yrpr rp'\`[ZE$Tv!HȬ;8k:LW.%R=W)1wlk+VdHLFḞ0Y)&ԼKhC)yjǖ;|Q9hK2ZBD*tQ̂J]ִ|YY \꽉HbMnTnZ!0CPDIyxrlRazd+gLB$a*':RQNx1(尠ŅP0Xy&laƨ2X#;gZ:)KSp  R,>Ts_c*+m#Ibv<"h{}SyDӒ<"EɢY:DknR1Ed\('#8o|gƭv&&9ƕJ(Ã֓fPl^wo)]JFAikWxL-.]f"#L,'?hnoH|T#’Oߞ|3~P^)C r(NT4oݱ(`-vU]& hߝJ(Ӏjiۺ>5ah:piRtPE=U4g:'?Eoߩi> U56]b-UkقX' gݐy #e A? Y]S +0CgGw>ंB/j")5aR1;oj );iAǓ,o8G;!J;-R/[Gb{AmPMCPYHDs$<*2^Sg %5+m6@q)*_2M 55~wfV" g0X/h A)5YLӶ5E׺NmTZXf|ŧ&X.+{A% 4Âhzp:0ϤW"kVP 8 $&Zq]$g!r!FV<z.=f ]jUșv9P$GޑN8EBPJD#Zh!4-Vp2ePJ&T=@!K(\:ii-1r]1g 00ᴷB2&Tlrj6rDOmP5*`#$E4) HoFy?TD$߅ 'T6?nPOc 4F1k;9ͤ0Fè$T7BgE-=F)7TS_0FߝJЭ|A6)%Sدv (c)_{(q]ܗs/4Zs.=ZasYi}9S:Po9L 凼B2GlDD_@y_[D#<'^{2 u %8q+&iBѲzbӉȡ0-Tq|A3ED(/ʢWGPI08j׆c J]dEM=xB@m QIVYaJXϱ~K sQF󜠏247H)KM z+zS_:IiH2fU7}8t}DUohFܗWtt&  0NQ@(H$k,w(9!x3K =E@xƭBƘBXmo< kB2| jBޭ_vs>zYJV@ .KA8TBy%$wp_޲9קxݔ 瑣Y zujh}nQg)O~ػFV_d_C^DB) G-g0W0'&) )g}*ڿOͷ.q9j> ۲_$!#qsu}u0uA.abJ<,]1AigY5L+x,xYC22!=. V@CNLyJЏ+ 0+s 9yO7foXL:@6e#ۦs3= mD_8rRpuuu`|{;.>$akPeL UbJD \\Q)t%pbt*&W[$2D=F38M5R9 4`Aд9xVceY/IUw*ӜHg 4`><$$u&t! ZAڐ suG*R"Om>1huu%u}j*5N`ɲ\ǦPOB!XvĮXj%ۻ )̩)'OEK1|E|"A %1[Ŕbdl `TRiCB*opERtW!5P r&8)Aę턭L| ea2YKP=&3:gs%Q!Ay9PZf92\Vsj\`1JLY&UJ$G+٠elh1\y|XX p 5Eݠ\pRO o/=Zr=:GyJ6%\R'K͔DÀpW19Ή9]֨Oly> QoW4^K*H&̾Ka ZU R2JX'ކ.笇\ZyS8G?<x9pEIjɍ;rx R@M~-}I"ьB5EGXeSL53)B[ӧO;L)k58v@'*eX 9r@iB9§qڲ_ά;~T6g_!p momM} ُBRK!Fn!Q#B@y9:cj! mKYyU\ֿ9w3D:/hx$;?,auA_M~ʗ}s}MCG-5~ffv>Ynr"?Hi˼jJ]ܥ݅O7@֒<< yrzڿ2\=p.!i>~.e:+Q*yW/6ȀewsUf?)B Xα[j.bu:L++VowC;wrq[q4Kk1ٌ4#Tzp~1+P]qo8z`KҖb,uG/@PjZc#C%hIOz-N;/󼩚:Zp>=+Hp34)3+k&qՅM.EJ&3s'}9IٴW( tB2pa1& Rt2u1py^glf@ʸť[Ɇrk>茾vmϷ.@)]]7`y㺠A !-#šhy[0#xƛU\}kop@f4_?|_[vdW{Y?`.g/n5dJx[|nQ )r@=ۜ8JVo#ۺe4L/xK;G}V0ڣ/lkI7+f/URnQ X]6g_l%ېA~Sdnf6ߖOf:3Z;,JsK(׌!)=|DOjah違08&Vtm*`2/uՊ rf9*~.*~]GtQPm:3S<$Ȅ/5i-)SRS"=XXhxXĆ+?|j0 \&n|^~7Kapu53/d9.]kn 0qf;.i _PלD{9O>+}nG %=./P@pFd1xK&NY}ٰ1ޮ5Y)է"ؼ v+ݯgɽ o3GqCj7[ZG#&~ &ߨ~}"d^j-o=:ղ ]zk]n=:Evi JTc(k%^i ff/xU\͍O? $%#(t O%oNC%]jdήæf ~ p3sZnBR㓴>k.g Q=|vH;3-Z x]C!qcP($sK NBC" AA X\`V{B&H1"KDB1y*8h-qq L.<9uӨ+w F$#_\wOWn9?pJwn_r gosY.ɷhoY'wϙ6 '٭ly9o7Wz>sכW_g,Iͻ؍ff >&˻y[$awQ)]., KAЪFQK ۛ^*MH1 #(%NKGΪmD-I Z&m*vf o`$W]ݻI7Db'9~AޭطG}G\)h+ɼ=:@zMo) 3'Hr'(>rqΟWB=HoހE}ۣ mo ԁо C%!)pm4AZKyKYh׫n t .lh~Ł&xyEC|rQlUj ČXUr)mX3XI51)ղ rM/`"(jN1< !%F]sF'xˆw48zrt9X5 \\+LkGGe+B%9IȝFȝQMHAx|AuѩY?[P }{3.wO|̠{ƍ K/[{ne7g_65V,Y E%l IICvٖ4i| COn =ecPtd쳧!q.c CG<)dB h|T«͠Q2J$; @" nǀs+u+Dsu/C5C;Gt8. Ơ_4>Ah-)4IIz^R=if!SNX3KCVs* 0Va`fGSY@G @rM)aͣ}rgVE1u [:#7ěihtM~]nܩd!LiӦr}Ir2وtVH`07W_d+ҫu|S zq㾫Hk~!$סvd]HHO s?DWD"]{ [Hi$AJtN\ t@#4Qf[q O$,{Dos0M=`sKjd63-T!/TӾw_;]L QE{#av*fk8j(r;d6}VS% , RV3iu)r{zXG,bǛKwG:lQgr"刳9*=-"ˡf_><~Y xƫ'J'QmNjz)rX"fPWwq-AXbWΠ.-LS8M=*선dѭ6q 2vQ#PG}c ˬT-0MX@o<-]HV٩ǟ=q_GItMj.Jk?>|EƷ oOI38_i~2s?sϟdɀbӓᨦR1Řq6o{oSӧnJP%{JgM0o*w7K ka,*D﷟ܺ-ץ^2GaQPz6i>Kbׯ./dluSm~Mͻ] T16wVP$>IצŝM_]^.+?gRkp!g]S@ 6Ѧު YR*fFW;Cs2|">Ds50췫mI2ŬHhJY޻՚ѕq))\hLD #=Un Za8PVqNx.>\MQyn}s&3KDۜXg Lph(OK$&?&V.hJ҈a iė]akKq27iq>n{g#j9iecr+D0&NP% Kɘ3ǁ(8IOwT 踦IR"UR [Azً@_(!]8f/p=w/"RPM;l_ŷA1qAq!7\F*٧ݠ0q-!{wVQAfT (CN\@=T$ fI LZ2>h/3࣍E];GtZ6jΘ M8q"Q9+*ՍDy.KY*q[1XBtݔ3)5nDE/?';e*:-II^y8ϞF紒Kƕ퇛oI y'~<7oSHɹZYfPPMm6.UVyC5vBwWP!1lE#WJ+ØyDvW~+fƠɅLWeuAq Ǐ;Q_J7ZùSHx5t=8/쁂8n\7+6r;i>q{uARAg k' pKՊGB8Z96I) ~vJ9E&沈Z%eP,(vJ6G0 { bW-v\AibvUz-PϺ87od)F/y#)(B2=fpԗ$fAA:~]&Vv٠;4YrZ0!bOnXZq=Z-]L唉dzɞ#ꮂEV(shxe#u{###r5"OeFMy2ymEy BkI 9+&0#13rFԙ"XL$E O`a65ߐ Aq# lAbAb~2FnҞc5iR<[nMf²S("(MpȼPRJUhak7J(ILGqWc۰¶8ZS@36ljm<ęa+ZڷX++Zs] Q[|3 Rzk^a{{tV@]U|>;HC 08%uDn(`MLYh <.Ӏ'K!e$aF%Z'Xp` 1&,K4!/ cq4MBa$^sH~SZ=D@n .iqn~}pH;D0].ZáqcUF#n ,@Eyഎ!!?TIêAz+bThBUd\$oIElvjʴB|'i!1j|>ak .|>Ds5oz\u;AY3GUgRh#zgpo1y#lPBoҲ[炓$s-P6a3-+#v_!&ݧE(enf;q~z{;I}^]·zՉ%@5kg g|]ݍgk޽W큞VAJH;٧%<+|ZGe"5NvD꙽7&s4̹i_/׻?R|r]gCQi^i:V F3E(4\+/˫WlV.-/ej)BF}Y ԃia>s ӛ52Zu?@F%4.r) T 6mLPwC*g;fe`+vӳt1 kL0H.]pFav"XqS4Kخ=h:e Gzp,-bX}r3˶@`zW gXf ,1H?sFǮS"Xn( ٷ4F*-i Y$wڀUH ԣ=UZ2JK=(9x]bևzӢ"3J2"MW ChF L+M-z$cEqfo7f`u|M#NCh+<:BwcH߱^gږȑ0C #vOtDOox. dvˤoUDhŖb88H VKXk"@p(_jA^%cl/`JmeK|xY f/,1=Xɘ@[q#TLDkx9C܇')SǏYz+-Yz<}s@VD'kzeOE yBjHTd'$2QQoAzMp* n0c*h܋[ˋV"aHth2gL/0iY=Y]$ɣK7ob sId=%4<>5X:6MJ*TT>)Tr,:̐懊2--,A޻#-߭oɨ y (2ʃX VaU̗Y)}Q1vDI"`;,!˄1\Ǩg11ft8HЅVA63ȷ~8ON#V 1sZ%NxYXK鸧54=oqC/|hOGnP `d wZ#A#p Fs`go =+@hYhXՑL򅧀ag@@HA}MN>ogRϝ@8!F`t0i- 0xLBLh귔ow7v N%\);S3X5L7FaX[ a7A N ][(Ԙ ;HnKcϰSGPx WzT:]YVSK taa#eHiCp7,~_nEAN&P2u :,OV߾ƯW R?B*]WJ`ϱe:G#a3`Q?*),m mj ${48:s4>s4_Z|:~3QBU,3f 9gCў g,x 8Dhj%=5noX9Gܧb}u?X`[雫齋?".4B*OfhQr<Aݏ0r*)V_ulO'1)hWv4fu|0'ѽ0~5Xϓx׳>)0HbpaSKF}ׇ4#'UR7-,].Y~x ÅÅ JxoXϋvbnܡm7QI>蒏 ݲ{1g&ƹ%h%nqE4B ]P?- )pQF'36Kd]g_/_sç_N|,Fq|_dfϻ2sE8"I2*v'ӆCTy#>xv 1(֛/5 yvgfszq- 4 a$ϳFs4[o5*{$[eUjNJIhPŔ5*(m#Xaz:*Q8CC^ΏG˶zWjϵ.t$?537㵯mWfR],v4aî_g%p4>f_gshgo5g7&a}Ye܋ţ9{1t~:?̻W] y6;|lEq>s瞯m mDx9p}e-xAqOu[滑\\fX=s̸Zk|w!LHcf 26vk|l߰gx >{gZ֣oz (/zO&,YSwFA{'Ȯg[&:щK)jq|wnt#W o>!2JHo鯳lJZ0"`&ڸP~qUej>GU{}|μtat^k`KWێkxu#gG:qhVX9U^Ѝe[g>j+X܏u#ֻ+#v'o̢+O.w#9.{&Gkq.u˝J)Su U]+qxFt,wڝ~nǎ&ATxx1Z{cAO XS8ctj*+ B*RV0h:-efY m[Ya8CgD6H_3Z5cޜPh EWrBh5is`դެPҜ5U$nVjv"/3B!E$37Bx8tⷄ>LeDztݍC\ҬD SnNz|[_1w}8RQ:1ϗNIk0EC2¶tŖX)+OFY EU)/A>J( z%JksH' *ewU-K$eQ]8.Dvrjv(5 !ԢjsUVRɡc RR<#{\0Vʥv2ʕw8^Рlfx4*\k4lxȠ[7VmߎĘL)ĨH)y◬i}b-'IA,ae՘1ĉDJwEG,?O~ S^s1Ty (k Id![ 썂 Ik,]4S&``fƅ|yza\a!Oo` QhAV;wTHE9A,.^B".I&\|Ʃ)C*o%.ު"8D?n"8q/IV=*kƮ/ 3/ɗEML{?z:2\-u]id;r\ &xr7E BBz> ~#zԯ_w眻@~rAݯ<W_6M43BD4,?Jsr%PÐ_\;AkN ;%3q$/#*PPdzb:ݏj0Jo8Ź[*lۨU0Rs^DH@+{А):%:k]8/&fܮnr)#ٿ2{! +QW6s0Ys49Y{UUyUefeeM&JpZ$`\*ѴGG-Dl"U* w[c7Q JO=;26{wנ(A6҅J4sY=$ "OXZ83SlQ`E}FBYp+P]`ӔJJ35M1Ii&X@He؊hВu#Wc.J\c$F7hĮ*xDoscCsC-Da)E4 *%-(FWBi!Qv1)(qTpZ̉B_> |&) [%~2TzJuj pF#*<TFBUs^> =&9Sf`QO3` ԽNG^3=*`ωB MEV TMQlc(0U^S"Q"D b#8VdR*:#\XŢP4Xf?a# \cRNӐjeG+8z`[(QH*7` 8 `I&@B1Nh&piTrq>&xI <(dHUk v{~تwzxٻ[s'~QhRKM( xu5Yѳ:f}E :gJ΋>ﯦ[ ę f/}TtsKPEi08w{ju#daRspoWqo0Y!/-~^ SI-b4Aa$&F>>&*I`YN.h2\ tNTy}w|WR_@~K56BvR5j!R`rÛ)I9o Ͽk_;iFbӻ Rϫ}~Ϋp9> :o{9rT}̂OI+N&0uS~hڝ\Ecd- ] S2\;ہ_uSb+X.&3_9"7CF4;!8WF{\&r nE}\J >By~Bo.ex]`z7zTsJgREK(Z>ܝ _Y*3֛^\(׀oOB;H_&K vA%l-t̲[M?zPK^*#xj+|ghlwCW%xGr9 328Bf)3202G< a5 ®ؿ(r 5J9ֿB$fW7$3uk7ȫA'?b ~'&vArYKMm[ rʾFǐ y'y";`KlBRVRnW2-4OQɌYS4=q̬ Wb]NVFJ~zV 4߄F!F4FԦѨ0(EQ' T>0 Veb=`I >Hr::jԀe0vX{cFXL y*R^Z3;>~Мj/ZE+ŕZegގӢClT!g+e~ΟJtH? ?_j"Z'ϟ>N*5˞;d@xyXNv>1#HՎ|]ׄrL_ uUGNq OtJU\H0I,Q7d=Wth%ZJ{[CpnKU=}|SlSwk~rCe(ww0PNuͨ8v\3j8^8^󲗣lF4XM3Ls@h͖z X ep;"boOAU0!Z˃ S-XneVQs%[Ȭ"'0`_8_,G L^1!u-A@Nj1n qauW ߿¸ٖigijͬG/z:y[ŷ?z/ 3ZT>8z'waT*Zyʩr0SRfBj^(',5Z<H+k_#OL`׹!13Ynl;-7^C?ӻ>8zNooN`7l?]ow=C}t|xwvtr룷o~=4uPqQvb]7ƿNY/`_Z` `u5G/K [lBӻߤBI`RCk?emӻoQؿ#<#LA~dqG0ss7P.iqm;V+?j70loEcm{t՝bs$a} ZČbj B8ӬN*K3g`u\˞ o\j2QMof;3diM/1/#w3fQ w{q_ ԸF5}KoV~!4 1l4_F 2<Msu7/|`rg8_8ɩpv N;?g@Y@#XC2C V@N`D?#&ɸˣ4D?Nֺp?vY!f_3(i=&hf\ 3J^ѿ9f{/"X`/3 !R,֎[Nuc)QT JYt CN1(%kʍB70ZZAh<MH+AaFLvh(֍+!̢j D4̫%&IןJxߐv1 j:A6|oBO`Vsa w{9Lӥ@<TV<w FQeZW*9aqzj ZTNnxfРPHcǢ*v31yYdvb\}ȹu^ X K(KԂ>DnQџ8dQegè`w^'%{ f/iz|Jsk1ؑW݇RDI>:1P &eg8$,JIIq;JOS\yvp-o9{ȿŦJ&MI/)JOe(s,z?/?\/>|-|~xNR݄ǸehW{p-ۯ8RY,Pl~SEQRLjc&ˤpU`xVhA-kM="Ú Oj!PAj k3 Rcc-%n]#bjcq@k[W "TeZ}bZ C$}v3 \=YVW.ߛN xٻOfO۲>{IJYӳt~g_&y k{?#FsSN3}"|Kװ7sս8;W_ײ)w70, eRvCOΛ6=e(QVmM,EIy)a1hE Ce/Q[c[r̴i4Yhz0^`ܰiu P̎cy4 Suft~5Uk&`5N<8!1&ѫ,d;Uҕ˰S,]N pkOQ h@ zC݃Rí;?]h|2i .^M{p3j;6mvWju IzVڑ>TY*+2\lf~{8+ ':B >1gAMcWUmGn5朗pNsӢj]TZAj1W$*eH:)^)nrZilZԀ=ieJ5W +Snt9wyTGZi:Bx{g?\Je%NY`ofK9ïƔ݇˻橿ņ&AZ2O.1tUs0ij뢦ᤤWhjB28H3f֌Ze(ZJ. RrKJ&,ܮ#pl女6YmIO>Iʮ4 `A>QJ@R K(HJЛzo3x_B>RIutXGvҡӪ˂@:)8DZc94XY&9 `?9ʹ]nqq؀y;ETBL;2X@ݭnԹ#;L J=9/Z!C]rȡ!{b7y poaJ7>01?L۬^>Q@>tqyY(I&i!+[sh+sޖ⥷ņhoB/کK!βIET_iG){wu_i!zq&o[ %Rlh'=IU!kW@-_)[kU 袈Q깢>S4eCNa Ucqr^?cשx]}؂apӏvv0^Vsg&bZ8YE;QvR O"d 5Y$Un+g"=nDwD*zmhr}dN2FN+Up*(U,:"4TRJ6Nu,!'?@`hTxrFR,PB*')q%T||r(g:M.=cFz'=;G)ԞOpEr%Lz 8YJj蔄zCna˟0sq*Ѯjp+8\n櫠{㨑wDT-r$CYp~ nGp:IZ;z:l'"NAepωs0"7 ^ٰB&%\?S^/8FKctG0#1˴mT❃ LF9;<=+6;2CbD@ZĐ#3^X{3n@IjK` C6}5[4hlո+dŜ1Fk'4֕{'.I<1ˌ7f*@l\mSr 8 epYn/0VD)L ^2a6$ DVϓѵ4Y+CL$OΑ@j~3-d`:'/HNLA1R}\볎E4vRI*SF|JY4aH <r)Erv'TJٲ+F{.[_p)G׶e֘-GsШ!f.A5?wږ=v1-Tkxoݬ.}{w~ڦtlŞc=Nܙʤ2K޻4U_>(BKj(bwZE,=7c(AjmUG@Qo1Gunf2 8[̞n^ȎdH7L|o*;vu"=ǽdn܄AsP^_s}pR#?םAFq4~/lLRT1,Y^ԕ@Ц$͗.R?ޔ疆Eu?~77_<:%OnӬ+aւ|x5gixp;XF籠gç nAKqnWN6Ϯerhŋ>s6yjvss׾-cR2S~fO魛ěKvI\[%pgSV ziqڴyv-&^q^UL=߽jM~w~!CS0Ll~iC?ܒ-XxN u}Լnc+};ّd 'td+eçLH(I(mGSv2?{GRW*ˇeֻw_rq5lYgYΌMg$!#_lY3$$(C<_?SOf,ڸ]<8ϫOKko餯ՌyzKFDnNoOǝX 1꧋ρ|T02&FN~SʳbW۾sXO>2']Q9=,? Éܠ@ [s\8 m8R ]%#@ eSǼ1UKAAA~U%%] gukUIZ0тml یf{⬦^qa!~mZگe%8\ ״\K!hiaQ-q8HfBmD $m b&DClgc"D#bn0Q0Gd\lw>#%5t_OOߞ=ǯ]ڶ]b{- Ē *rKw@D5^ʼnYC?ڟJ1ދbf 4oh#é/fdbV>Ώ` zyu]bǠ CuG{a?zTo>6XFZ>iŚȀ_1y0#=$>+1Z80_fi|1ZIÑr4Rz;#-].[:c,(oY:"dxrn,IOv[LkՉX+&ɺ>M_'L]*;~)7L'fׄF2[-Ez#Fi{9?TFMyXJ|^:G d\n_ݎlc&Tv>*BS;E!kʁa ̮d$ jhmA.N5ݗtk/DeɪgY[,8y 61ݱH2[g#)nM`ILo'׻w=|noNg7Gά>x6<<٢r7d  `,3[Y"irߞ})fgl5Y쳔VswX!`Ih{Z^ <"/O+͋ڻM"&Gkwڞ)򛒓 ʢ4p.oEi(Tky-迪b('|FU_TNS+gן.leXzow*}ѿ}^6 L5^2 a}{tܰ$ ]}6"+>+z΅|mQb z˺ K[ѭfcѝx Њ~Ecy"_e F!TsX.0s*aӾLUu J 0'yfQLp0U( ~筷giOw~3b}.o~ZT\Lv0vc=b| zymӀۅT=wt=kp?qQ0%Wy.=PU ꤏ"P%$eo Tƀ ST2~*dGGNe3p20bwB Mӆ7Qi;d=LĶ<=c#v@fEGkm7S\0R{fSkgw2RܒWBj*J9\NB chߪvJPjn[=i:"S}[ s6ƓJ_aɁ$kC!8mvx!m -64 UlQ0˴ӎ%zLXhpRGi!SRFQH$*ʗ}{Y1 +n-lUI/<O2BZ3 R 'uՔZ!(kJ+J!\ WkUV0,+;vbX) P/ ݚ`Ĝ)s`H<4tKm$#&0pdbk.ݾ08OJ+Yl>Lk˶й|V 7.S}vAW9&ةY8mqqB28^)MYM*Ll3NQ.@N{%#?Af`1^s 9DԼ6&@cby%MS _{Onm#c{%rߢ&.g.s著` zn]HkԒa^<\wq*x=gQqsCdJ _O1\Y!ힲh}\+2d5SFs{JQW*=S*_f*8Y,s%,kdxқ-ocZ3#b͌fJP)TlA?xh~8s"37}q`I[6g 3'YOE0C$}36`ț8,VN翼#pZwVwMt"M'n|$~dz#r:n^xI\+9c#º*^bZό TVʣ -飦 +qv~y˚9]Q0tGoٰ[oN=\p2F8Ļ@`9ȱ:=oR!J΋ *ӥ* RBHDb $JC ΥЯM!BJ9c+"rK]1٠en(K^de ~팠qa/2N4?DECM\ gיW28YBa|!֮*+pJeRTE:@Ŀռ!xlv]=,W:ћk h8,Գ ˢd1(;O&ߐHku/#S&%PΫLsB[c$N=ۯ1(,^9UZ/m1-趌^z:,P7 ^Q^ޱM%b.;hړ%ň>u[";n|A35^a_e,EvVO> 0\&+ړ@232*ŝF=3C/4SefkTCF>,ߤIlċŖmluP2/FGgKQEv(Tr$ %dLuNjE2Y!kM=dK`\ xD~r  NU1 +n a{m;5^R);kfںC 6"Xju'mleCLvĄ#%Lq/xYC&7x_/$dy c 6c6yAևI l!LpsRbH&_n?`ֆ)݆ q {7vR6<d@IvE[^`\I)f5|B yʢ FXSR]PDJrei_jV+RCmX|Zt^Q89Z2BHa榸p8AI GM#?p̴"6jFDc8l0B"s%L[isk5Ӫ,KuUf^ т>Ľ R?99\ ҅~`||]" 6f/;wthE<};$]HEZtdB{z5a m1,:@f LʗbEVHO #8sCEeL[mhnVlmz[ [h&>H[&7\]]0RmQфޕ5q$ҁR|Z#c$e4 "ٰԏPa5a~j}!TdMZ=wr vH {@uD]Ӊ:>8oD.+ /;ݺ]xIt`JEV ,qI4re[mVc3)م15DDx_f7tR2Jy+,LU.Yeٹ|5ixB x/hܫ*[ Gx ܇R;,eqzEUAdKY 䪖.đcGY~?TBeԛH5E:)EFN;ī+RJLңHu&CcvuoU,:VtKMPc{xUc^{Q d0%RA sQhRg'&;pz8/˶x2=( ۲O}[v0rMXc+@uBAGEB.7cQ8=z?̻̱JFCTjaG.>^caK@$!{~x4VyA)qY eݧ'!R~\c0f1cRZ^Fm U\cb c5as1ӜA[ƇDmU"*!(%Lϓ6>Ò*3z4b?pW/ "fw1V AQ1RdJC,dFPɗWPP+#gQFFJZ(|)^G-xGnS4 LϖߞxDZJE%M"B>xj}|:/fod=_M^֣&o>^ +qkjنRNQ i$g=b{ [p]Vf\!C a|4З;a꼍˷GA5X>4Z+9cQXqȞ.϶xFE [cm XH>*tobgSoѰ%Kq)=@}]I-NW6(ZĨO:fEfnȯEPkX~w':.htiG\}Qt:m h-/reqy k2|$)/>ә{qN n^PZPH8\k$IF&[G69$K8鹳dPQ(){q'ݷpB.j18&1` DJ{A!$X ĉ6ɑm$&L7 ]?&z4odddQBv)Qea`u@兊"h#g]3$gh#x1êwi]UA$b|\UIIy~`0oskd LK>cj!fV;yjR,/>M5.UurbWx\,gtx^,Yp_ޞ.LgY.oa-Etv:ݎ^e'̿jsɋF!ꋳslu8 hP%Ռ"D_$yyӫ \QIujE-H~Vs6Md{=pxI%B2meTJoQZX=J Z84ZC)L ʕcy纻>7ٓhR20ZvO0W-A; H汃;oHAUl5>Lr̡{i,j(Rt,1omWs"*i9GE1L.Ed>Ø<ӿc/rFXcE44^mĎ31-?V1Du 6[ij a= rr-}gwyui2*jhBp ֐CyQ4zAv]] YޛXQdcT$ؑ~l\=KʩFp6oq/J=/h>RB2ΙRi8 w-k[N ƃFB1 P)?ol.-rm:R4!HcX굫UJ]RjΎ H¥<2^ڊP0rWT\o/@|.lBR9-/$z!hg$R蔬(q@ҿq/$|I]yOK$!S>*9 DODLYN r?"*Jc'20\ZYv(B~A'Ը8MDPDIy%eq-`m0 좃X|)N|h)2&#>9^&-׭ { s@F2bU &14%m[/,WTQym@Iey#e ˆޫrԣ"a$ bLjY^-Ԡfk@TQh3`=Dh2>Hn)+֢ǦK$ W/qj_H%Pj%hH-9:Ί.pM Jg)wFP*;ͧȶ5WH{wLA -2>+m aϸIɁH7)q_` i0s+‹hjaE%Wd(~rmi&9,Ҕ"P4BsGN{ʹMddZ 2  imFH)CDBl?& A3fפ0$F83Tb[ JQ28/@8|3BއD˞M 7BVF ҀΖe?XYQJk]^LReo~W4$ k~J=!} \ zg.S4O e.Lb-Ja*QЗu!f@s1X#\*[h2ckWTҀyNwعY~s♫O(-#f|7̻'co}۰<- >ț>z2-e~HOggɩD?|K遃!Z`~?ZwQ6%0r ݇ƄaƵo|)8A"eE=RKKr{d %ixtۖ1Oh$(V45|v̓mQBaFPH @L6&GCx"c$֢$#V{޶)}@ =(^Y?e.Jmױ" Kwo\Eޢp$epQ<$thW+_TU^T)ăQA}S43ܭp׷0|< im{߫l=Ik%f K9gC;Nr F kݳN~lf? ;\כvLLpwUxl%1p6gnn﮿z6PVT#|;2Wwd|ZZ}Xkzt)DҫD)<տK{?tcrW<漸܀W*y2!8b/A ^تڭs[&>T2B Ӷ.93̊LHe[pw#p`{rv5,L5 ?/KS5DQͬƆ;J !>%X%PGS4b/HMM*)N5T֠x9SE@y3?tQ\S+$&$3$JZ"  ]NNO+`eգe2@`U!8I ʍˈseF$3-,| ;%et ѸI(2| X&\)!*`AOdQy(_̥R`I5%Y&\c CN(B dok텮c-Q9U'1d0K 8*>OwÛk{~=s(?pOʉr+}J_{(Wr1W>Jc%K3%P(HŸ9EܕmL9R<,1ѧ]1'LpaA"*Eq*UFo1F@T˗#AY% P|$q̅$*Zy2&c TAX={~QR!rx-^7NUdo_TA `I`HOK"u犓g^1 'fKn3y3{r)y< *bXxӸuvrQX ys<[s偶C sԳLn [7lrϋ̯n[ޭ߾y5\i#rbb-/o{ a.' Xכt^HK"a*?(?x懭t/Ǡfh~p+`obyߝyJpt5tԇ4HRkc/iZ~w^].j}u-5? >m{zi}֓gg\I*PߜOf&k{;}L qvԊWdj7-7,{L4 r+T` 7fS2ZG~j (,W+Toy9čjb @g#)5gu8gJd\-]~iv>~^ٲyjǃ0z6W5g?)P칬0]ou>3ugs)D=I"a">p}2"& Ec IHk _,%}d)}2K®;tӠ|*Z=_ z<0Z[΋/j_Jy^,L xd;UZ"mJU:B'f4Y^)A:!,QsuV 1zU-|]:qQܥ~%N⥡%H~}P5 8t^PLJ #:E" O0 L`s]׃Jt-M+'DV!  Kf. f@`PCf:z'3TNV)87YO»|Q}BaJy\,]̻5R*&8g͂#LTl%ypK P#'}ć9:OnXS`sjɩQ1,P*=>ݸݓ>=L>O7TM\ lT[{FaFEjKMe)hA[e6l/jW4e,Cm,2S`"*7JIrDU\fRfDEfpDӽ6s=,Fi=l5wdգX zz 0BʗaNff@9 N i%, E!J*%Ma#Sv7p.#mOD dqTXXxGYI7;bTA"2.fqQH3;W84 H3E3,^co۠TIt>AP{ȑL;i8l Qwݯ$AV9_6OL~r[6Cf;Zn,99roIgK(X>a#W:4.n++њY>] xO`nGlǸFaզ!aã5Mv&pGb;=5(_bs8PsiJ:F#-(Ua(!yT_09I\0\c X j^שŜQͩݩV#Ɛ+Gu%\u" ;վ:Ij Ia+5O+(~Ozp7[[y/[5v+ U^Xy|[d6[U (?H}GEй<_UxCT_>Pv_Z?Ⱥt;~zƙ6TS<~v(k4+W$R=Y-!v< dͺZ>4+W=Φuc:Ѻ DuRcݎE[:xduCCrmT`Gg.g`g)=;u~'w=9 0aPOKbiz1޻a9%kMv>}Z8U/k H$y -"Q"yWP4?PR&/LɈ]PpNA(}L_#.p "}2 %azٍ In|Eu_웲I涊O6IsVdWL:h%_.>\t!?EZ4_[|Ju 1CQ`Y]Exq cX/ CN };3ef-K6uh%C !#ʨv: ϫlFQU •yaAr5KXZ 0 x $h8f,ÄkqKp-5!Yfȝ#D ]J[d\pTGNx4 `Ս&La΀H/Ak[s`(5\(-qhIrHr**+,Gin89v8fA(g@ r"%F`=2R DQO Dkh fAyeas<:bPY$GQB[<4E`ܫG޶AOlb@v^J b#-u(1eyFbyV,{62<R "C2aD& :aKF6infzu\/k=Zjmx^瓿Alu[C\}^[13ZU!Sj'Jst xۥx Lutwπ=necju.ʭS2lF MIҷ{Z:SMMgY@`~L8ֈ 8.݀xX@jYᡚ /v[4K6kdw1߮ɘݩ } RI=@ķ?HaPz4#8EKzZ R ktD(EH5!"'-إk. RI0TRl'8j%yW$Ԧj\<7NP)k7O֥vbz\~ITpHYN;b*ǎ&j4r,8!r U4I{M1O DuRcݎw$eh-_HuCCrMҩZRF MԒɎiȃ wDa1 Ti0Ӥ UNIQ? ~wSX`X}=(>"bu.d-t+@|(z%l!Պ({wyڹ*ƻ+tX:+lxD )EuDnF̡Ct(jSP')L bȱ P~v G&KXmCU#dN]0NW |Ejr{O u$X,r]&R,\j>e! MNт27ҐRX}PJ~#wFL?竫uS6Ic_|>qmSƝ¾Hx-K?{㸍`#7eHdg ]igܶcg2俟|kS(=`[GVdU Efq=D]K}xɯ>^Wupy)uS"&)tB5x m0Y╍wSG wr7o%޷dq\k+Tr 1͸P\IxLlq.bM8%1 %:}0: )Qi/,31LAKP\ީVP{2$@_1 =i_)wvYW ,?sjq͂Nޠa6듄Y8;x CO*N~9U|8c7-enŅ&拙mmq$k!c39e9G#Ć & XIcs%MbbDuù~As8sA (7ô쐍,@@9w9IqA#Aa4[f9QiƠǺD[ް*7|cRa8w)}TIjr}\9GB"  _ _Dzbat{eY69l(Z:M |ei\3i0\K R;47،Ɲ~{u 2ŠInҠ@"rsR$_I+(p\}XR#-:@WpB9&m,!;)0m5Ep nkr'վ-Z]$2Nª9E8[)"{O9}B=]ްzT):'}-HW#IK1IJNV LP{Q7l;+“\lžK:[56}ms~C.ղW7Pog)!S[W-=- "oAxņ)Q^a^ Hy+rX>s.1N={3~q Fzu7k׈in8cpysCջ*QUG|ONq1gT0mc6 v\6vv  d6OD# :!/`a(;cK^{O[8vJE#\d5v`Ԃi'HU o٧ uij; ( H( j}bאZ׫6TEhD*BkU } +b=J ͮ\ޫ(+*$=$4 u)!Ȏ(@ AzAҊ\6cMmD'YƉW2urǨ0nzyL"Tu:_4v-<'_`ixZ( WtwyKao '6V#q2-zjtI| P4|k$_A?ب,-7㒽 `Pv8Ͻ"xD%136+R_7*,L$iK(<w288(Ɠ]XmX8:u8l#nѺ;΍' J˫f|댖رu}P{p1@b.9Vq+``oYHWQٟ>ξ%q*㢡qBp [@`MkDcq ]P,(,XD(SUTRJ>6, f1J$Ja {F0iJE&EERc3GK|j(TN.>l>>4NQ"AW(I41T3!(),$c$~7|T꒖cMvսϳSn4󧧸ذ89.vcQ˳}]y:NkXס^$ê޽"Ϥ7sJ-3H>-[̶p`_ǯe_Z!IbTmR,(6$?v"gMV1?@=.?g:ZNг_-hbN~_דNG;dx,OUb< "8gB)Å)J I*3a"jGw]_8WHs}4ShJ{ n*7@;"\ H/Bd׫n~URS%W}kW! lTU\}8W/nb4T"ݘv$0u?FZ/4-V1R=oxGԯeը}K\0Š/f}x:&\X&iDBXS,Q-ں9nzS>7R3ջ0})꿛-"aq,uk.},xR0\9 '_诅dgCl{?bDU&z|6ۛZ`}?603hv>_"n=du5.EeRkS5-7+]9ge)gGŸRK?G^mДZvߧOu {`a@3"fQ'ޱѝEsxsYSc#LKeûDȣx [|,fqyl3l')jd~x~L]Ώ›So7b@C¤L4}xʹs,>>kJ2]Ԏ=ُ˻vֆ~*>\t@s|&\PqEXSpW h:T B̙40S[r]^7?Q8)ZLM9c)Nd8g<hF_쭪:R'oӌ=ewBĽ>e MBX L)A % Ncěۜxk~I$i){. QLګN?U,]#.ciP֓L{=wCN߼oOO++fu=|(F~>9_rzŶ@iUY<3;Km BMIZ0b#PuWrp?cf|ܹ=А]y>;9ǝ]3~n[ˮڟ/߼o,D{ޯH"k20UH/Ra+/ QJ'ngFry[ MyHU\½忐#frݑ,p2 lAٳ OD/`icY||thۣco XPO+_ v9)ۉheTS~JRj[C&a9ꈌNGPy!Z¢'ѐbspV+`zȒEHDȌp I( E!KMf(VL@/J'|Y`OiҊ_;_ʝG3}GׅqRh'U\9:ؓC߽$}|u"-c7sdVù(7M i+sRf0+k; hI;σ$J+xV<::cV#%kpJ%, Hه9^[px8D7}a;}d 3 H۳v8m.e;nX{s:ڥ&NG3#:v:ZHJ?{ȍENd>3EpvOd}B`؉my%y.g[ݒZ6[}<#`HfbKYtt-7s3Ve12ZtWR%wtn{2'5^b9\1M1ů7i_Bʴ"jʇyL2[.DCV@hJjĒ噶6τ4=HvxdR˱i#)>`&ƓObREfѐ̜*7P_?^7s_"2C.I7(c(1`ł_;KY04 ![6_YC 7[ؓИ uќ}MlPlNV:W8dU/HjdBA;լ빠H#(z^3f }%&zH'>4BT7ۧ Hn?Y K'?9 AAj#NRFkF|}a=RacH'S-t\hҧj׷޿+%w+hq`A![o`-xYͼ *siN -ljL+uJxRRҹazbQ,RD% 6}Ki?C?6(汕SJѷth,[٭v ϯѹ}gpuE|(6&$?T˟^fWj.  nj[)P=1)aL=Yba8ҍgSɄ7?oy̙yV?j;Лna(-@]pB9嘣e jq9ߐĦ46FbkC+I‡-$O jo q$Px8@Py\AxFYlTh\Ntrv{ 2#WKhFZ&j&jR= BBNM(T@BBQЖ eɈ+ps.8!YJ@s:XG[d 7uȱ6B)Pp4Ou/ *7t17(vY@|a  Pk!Ծ*jpP[ L-9h= /'( BIR9j{J.nT; kHoF/PYTo:\uv9ل~E|ݗju`)rjiXJ\j\دVpiB1R|)!P5w! ⴐ@) h-VЧ]tr`, }aYor" Z9ۿcjc2 O64$|"dYJٔռ b9UqЭ})'y)dW5{E#GsаSn.~ }YM#o1omZԊ7&P'* on;Ht'atH۰)dΣ<ѝ2'6h@B8BH]=F%MN!P)U P+t .BoKt@ِ:iR\/Ɔ e(tuG}iiσx`8 Ji @b 1i# 6~ۄhT<;pIᲠrQ(-\:Hh:P(cOpQpun|'2Y`-SܨA頕n~/@FpLҗ0Z崮V^/.?JwB*&x@tWur }$#G8-f$Iel̉vb[K<㉦5cä`2eo1 N &J}_`G.b%䀹ۨMݻ`7"Ϝ6CwNfhL |1\!qy|]e72S@Z*E"V4; a82W"g~(vA&i[ɉa;<_]s<]NR}9g-&HeCwKV{G +!hS˼P2 S MRX\W f j7[$&A|QyohSs]NQLܴ`D#9~ 4 \w1rNUo)HsilMzW"{d96#m(H h ߃~O&|CD_r[~rB(ͅZ5JzMH~(/@G7bv\ȸdd_V SMLxlpmz\GƼ=~'k54J 7#Z/.rxģcFj]Of=ϱn9h~Kz\{`KsfKL#Yym>,ѱlǙA߅h9`ѯaĴ@}B)o ˡ\^qq9Kՙ }O+.Fe>a&z!q?h++A{a,6"|ZE_@+1z+pe-T7 PwW8@ | w V iNsr )8I%Vp#@!ϵ/j _)'$s.KCxkN iT 2&2ŽTNJ&[8XMoFʀeaXlv疵f bŷQpc;_{]Wu1fLwRc,)z_a^~|qDG&>NO? FQ9A񕁍V:nFoaQa2<# EtNvՠ"& Jؑh_'51K1SPA"ut|!"}{l\w{ܹWլisիWC/$a5HAa,5OƧ?{n쏰.g޻@ 9ʅ%ziLtoهn{# z 1+-8{ؒ_ (xpswNni2ϴ'*WȭpV"Sm{p+ +^u蒹7zy?\&z'o^3~^u W jn|ZqQ=qdK [88,UJ7S /rlI2Ia.gNTs) 2enȥDC61XYɂhGZ6_k9T`5LP41^s)Ֆ@18N.q|zp֓QSG좪U::K.,jF ۃӳ;9h2u&L=>7&6F3JjS6OSfuk59-^hqLbf-z4P~!20B2|>];{[ԚJdBI$ 9 j~$C]~es9Shԇ9[,gu+!oM7P]ԡ =$nOBp&uw!o\3*Qr-k*i?;_:tt9;v"(]xn+gn. eK㙹?+݇6^9̘q*Tf 2@K' ycCۗdr`2\m[߫\8>y*npk?qF_.2ěx# ƎcΈQC `%0d(L u4 mɩ'&΁fGg;\ig,siX"lL@ =|8a (X ¹SeZ/8!9kEUlk%v9OE3F$ v4p^Ni&ejRI6߼5FS*5le\&13 JqֿTBPMa6F 1CZ+A0IV?hDZ"d͏Q[RE~SDcM$/i1r^w;B+6jKvD"b, 0t[LH)`=-! %.fChOB2T*$#%4⨒FXTyƠMLG➎/ai"Kֳk59ŏx;M 1+$ Ab/+'*`P((LFsF2ntNT.4\BSP[(T6yJ4)Zos}gE9 bՕH̄]_mr&h/'n-U4o{}%D> w$RCaI kDV7[0q*D֛|w<hkG7r}V[mSh^/hFk5waRBms7 |0 yuu&v/Co38,* Ͷ^pȳC+צ0ña=u~d0|; (; k/\0tRNP'ޝi\wR}]=);y&XׅHOʲC)(lIIBʔeY0ٻFn%WZCoEd8A!/9+IF98}-j]lwղ5#$@KfWU,~U,eAXĝD-3*^D3xLS)N/ %?M.若fCF05\^:吃033O Z$ a3:,GZA46NIҚ7,mɖ|To^a S@;g=3+ioId׺%B9>+NvgԓZ DY5ꌅF/"AHhzCd`-\)ym觔ivUWkhk.GA^+]?JoT΢@g[b|a QA{BG(kjs?F6N08i<;iK6R}ZxvM V>ѫeRO*PƇ7>7# Lr+7B3F?д6Ҋ,!j \ZG?Z)_ v8 0Qt$OQT?ʿgd1_&_l3qCvmVwu]'X5" k m}6jJe#x-B i#]ZDFh --YTEYlFs֊"9-s`\̦sdt Vܚ$lJ+O}QUfAR/wyt@^CnY*I]ZtrT|43"44E-$Dǵ{^HM#}^B=Bd3Yvk!b4c "޹P!V\6;Z ZҲ"93Fq 5bb*%G=ЌNRûor Zsz# &(ʙ7IT[^h~\4&vWR3Ccd+Z$@B&ɐ+T' &h^MrA5ّ0jA kruAh?N1ۈ@\73-x1B &"nqY{nsK1׆A]5ҵ/(j>|h?wW]ϯ~ʼnhu? )5AP{-> Ʊ53?֯ [{`}{!{J5B4 *"{-:#/)IK\C5P-*kF?ExNrW*Mvz*"ís+@Α\Tfgzkب 2GsryK.5)!k0bu>cow `q0sCaW`U,-O]Rё~b%҂Y1'JP2V#6…FR/B:PBg/> Ii ^|õ&YikX%EЙJgm #ǓԲ6mBP45j=Q|Pjl$n,IK-uYeV'z4=dPv\F[Q_{2+}8?|"ḱDB 0mG<֕hEeEr{cFa'm|N,`*|=؞9#8}HƎtCG=(4Hdid>ow o ѫSVN\%p) A3Vx4IkgҕsX ՜. q0Юm0> gǰVK;c8g.pgfɖ'+6(8PF&7[E/ + 'S~"n*~kw.GLٞkmo7Z'6agC 9@CĬgNo:uрQ0p-A?S~sCr-TQw)m8RidG_WA1s6XR^` (^^kJkR N;0$?f9W!- ~D(<KyM۰տilkȿ\}e^4v{ id-.$Q{S:ǻT0YHz~U[?v][\u1z6JulG!mi?ii2/zrP|-09kWMR+ыhFO\vs `nMCv˙ZV:Xӧyšmc`m@(K4ik? ;eaeum۰>z_{1Dᷔ#6bᵝҾ~ VH/Vix;NMtZ ]2aq_󊻄_U @1t6 {h ?oI-$ol~(ۻn?H.|vi.C_r $S*(u_߇ܪid{wCr˜+qxe7.=oi (I%If_Я+(חt1QP99ͤ|U_*! ږפ;@W 6/NmSLԾ|.`  ̺t6ЭID83&Z 7,1 *)4Nf/}i̕샙q۸EFZm^ck3s}1-iؒG}|Iui]>wK%AOÙOwvB\)'Y6VU--LmQgg,4蓱*0Ϙ`$F%/ဈ V<4uf$4R@kq"I$y f3JEj*ט eOAaYژ#sL(D t&E%pERW*DLΧݹO]M$pgQIO~~w?}].߽`ƌh,Jz~(NhF߸,ߤ0*)k.cT $YOԉTSloi YkL ʴbJ%YE@xbfryI(Re˛/smskw&71Y&??6Km˹iCn=zZ7uVf=_kAK 6 nZNt\?R={!ZB`z{;{-?LRkz-:y!kRJq\ZhL3V*Oclacz&L#)l%5<@N*KH\C+Aaji8^\B@KIa 5; %70ޝ7൴|Mz=;o vyS!ܷY[0;UdQҷЊN;oKcTmrq>&c>'~neȒ[_4b=o; :eBD%iT\e>_f=q_-ޣkࣛ=ȷb\0Izdm нTb /7\ 3 Y^ravdbe&)pmӵuuUu-5q2pEɈq [zY…h8-hYrk해|Q"g5A8o#2i8hh='3F+9a#K7)Ml&jT)rХTiSC 4D?:cI%/v:DݾœXK0xRd אSqlFN#gFLV۠,\ R]I{oC\׌.GRۖI+Fdv0b.&٧UvVMϟן~Ϛ^i0ICI[qu١Ҽ%L"D\3v b\ɴ .Sk,7vL@i;82.ĴΎ$fF9IIۯ1h|P 祘: ijzG݋)٢N{@3((Wsz5,!h0D*J zn1@pDe!밴ndCI* ;m}k*'Dz$LfoE=2 u˲zdTkiuh<.?NmF%%z14߄xP7(o%\Tػkp €lS-Ap;ۊskd |h-ӵY[ R 4(R>Urhz}0#ҕQI\Ɏ+s%}(YtnrK29|}?/IS|uK=Z. a[k-?.wu_]cݢEzAx>xZ骟Td|:c̭=8~/x.;.6_S>L`J |׬\^Lv$!߸6)G^ H '~؜ib OE.ޠ~x"F~f xyU',ofI]-..cHv&ߞւ8Fd<4 HA˿@,ND4w^-ekч$?LIV@tg7?]zp'77QF=Ndj-c/E~IE= ˒.p6DˊF=X-0KTЎ{ӵŮ?L 8("ir#y# Un$:h  &H EXA$F*iJ˹Lq%z{!S1^n#u;p7_U%_Ị2KlTw|܂A:$x_]u¥G?_jno'WjdPV.Dg\@`Cq;"jJ~:TRwshFf}dy Ѷpb(;|G2nwR8$+ua.TGuq`H;bWb#9I<c*jq:\?KϬf- R3v3[l| 58qg%<(+5hq⡪ r䟖1{o)Njl圈 PsiCFIPAJRsX/2^b-nHąPE4E?oWiJU`7\2#X@T8]*+AJJeԄFG6.b9{DYvUB"JWKUuN֋oI͔xf0HKK[LўY26KrTƤhb%{GT`a;&7D2CƑ i'HYJm+c+jmT]SB;!ȴ@ ZXTImvԚPg ="juH@oDZS|[2LrS9M:Oykc-^(,5lGS3J3D W E..Z3n\`Lu^RG)X$ZbH Gԃess1ES+ܿ_~5ϙn(?Gj:MA i+mu{'Ƽ`J1~2> Ʀ2(MHk}Y0Fe' L_c+%Ѡz> _ ﮉS$(P1X= . w=y|Ux#zPjJ=^"^I*0\sI._>=8 Q-JLq}.Z:S2Zغ'C'x PW :oG1Y+qoMLŘS!}zobt:t[M}w-nZcd&mTq{(C+&WyJ:V+]|Nb| HPU 젒fygRyfPUsӬ r*,V .u橈~ڎS-܌V =+Nz8┳{qcHo ֭+N~x`U+E][x!N8&vg5|4)_x*]x`@[X1_\ f@PW/ !C ;son2)\wb\3>3r5>Ճ18ԀxlEs2='!D)E-ɘo hũFˠ%qLp:2ɫ[ustlq`H\M mb肦- ؒ\A0#kG^-I:U-H! dZKѾ-t h);E,R-`zB7aޙ0FDSQPEMȝޒL4ӜEO~}ŵttkMoYIMw>.fbdգ4 QcTU93M w;Rܓ2I .?mj5庲:7<` !6 %HMX"]_ gn6ol`߬ftwـxC,nM%*Me 02V /ӷw e "P-#F,Kde{ָwQ, !&rQ3IPə7xAVdL$hSM>(QM:G&skк,e̍u "QT*8aRe'#ߊ]DL:XW9(Zvgbǂ}MA] }R1PD=Y#!KŤsFӗR:YJ2b& DuI6ӵZs{@&a@Nl1S5g՛ ȲIҝ49KORT*5>R%2f,u2k')NmzD6!A5v퐓O813ɰxY1C/vK5dTge&r%,!xB(He=EI52 Hԙ󆽃:UfO8 ȳff'qCVijPh}x{yȳWh(m2}MwlzpByZT ։'@@JSm1ͮ򳈳PZXhzIv\7:{QyBou(py pk _g-nw]Gq͇#Zqa FEAHTU+p^W}Z`&UӉxfu"!ZL-1Z()ZL:WbG⠖Md0S\A{C!DZrRRBTCFָgp3z* @YgkukU8GϦiB 5M 8dmau|}=|~|{9߀nZR6®9rPKeg%%h`zʻQW8fPHk)t%gCcCCSoN ^ ͣ}."qGj'%:P笅Fg̮ZAL*! 4ۍ^}Zg-4䀓v68O#:G߆O\V _F3N|pZp !׬RY(R1}m[Xn@CȜUr #TȢ,=#H&rzyL toD@'ELS 6L*3D T4)5C2Ưss|P *.O8si/jdhhR"FAK $Jezu_1U EQvҢ@ Ly:e)qb;8^ʉ^d='kȏ("Y:L }n}g N ^ׄ ZkZNkC[WJQ#}%ze!GM杯eI:mMd{dFdCd(u}|Ź9B[С_8OXo+b[ntӅ3hqE#:P_ OiR,Lt5 C>–Ddi"Vb:ywl&2n[G`TD$Kρ08A.$5n|h޳Oյd"g5LyoWwW[m@߫;ǗG2;Wkys$ˌN %Ȋb (m (i_>wx|q2}ƒ,Ft &%;/Aގ$YS>xYcڤjgML&`5 ;c%[fKg9F9^钑msO4]%t߉f"aG i]$i kmÚڷw8Ǐ'8Pd;X|R6@xd8tsB l\TK5)Z3{7mf~ <~cnMF,9h ~t.H}=y~#5,E7aakqޒGkSQ?\n&Τ| >R%1j%7Yw]>=}wg 7AKOBb[<{vV/8pw|s[r$6C8oW\cs.ס~~pv32-$T;qޞNoFI@O jGIA6rhϗu0 n+̰t#қ7 `s7½|χi$tG+}roVƋ6]2g.N呟#&Ң Tוj2j %[e K6E ;٬Mb>oh3$#3s;9#jՎ*H8k q~Chc›4; A$$dx }}k83y!fweQ߂ K(1x x9fXȤDtƑN <ƵI>{/M/f>FȢB:F":|O8{FEIV6#,gnr׷֮Mq[G2&'vZ8S(H{o3R: '@&#cg2B82ԚhfNt@j5B8=~ `qAk`> >Oy*-p/OiXg!}OPq5K@m$j:N΋GָU>X ?fr==2ܸHUI jJ8A1J 5;ڽ;d-];ƥvWʢ9eD0;Ǘ}:R ,(Am:OoqsX/}:czb3hP $.wWΩT=:lz{;K>Zo޽cSg\؛qlMcѵ%3@î6?6]ee=6VX+%=ܷsܬu`]n8YVE][wuB2D{(a?(*ekHJQ4:+[2*']듸r͹L<&C-,ɂs|`Q>s f.4`5fHG^ʶoCRg;2桪uQaf/kk0Kv82C$2"iJlIgq1KeؒB\q4".8Ka' ,%R fG<#l1u~5ݗb/?t/<}g`H=z5A!jL+ipP\M3d?PqtߛbdZ7.+V[U[! Tײ-[ƥ/hS\p͐Uj|he2rPSQPIA_EeڃK c&%SryzGD!g5~糨^^IW/hlN{ ?F[~x%Ҥ^1TMz ߮)P$o-xO,j[3/k$y$콤t<;ΟNoސ ]}YJt!qvݽ!%$uxU-ɀuoVރڏqjڳ2HV6vG4PC2[$ %Xt$Y)垅|-^j_XiEU[ĩ]9_J=>P,} ^b[hXԒXkI-awEo`]m;$:уdtJ!|NV{.XEV[$Ԏ/{t,6xxz,8|eΔ`/bu|jҾTNev{Ϭ93Ti~vOx%8z{iNizVo8 4HJ>dpYKl{}|-ɩaس-+ 0TS`qaŨ`|F7%>y ߔ|;Y ,NzM2rʻj&n1y_꠻N.}B ~ūs2QjkI Dt ƔM^_ZW9 -\"z@-+5t&`UgIGYzCآ?kb]ݡvķlM&/,`(Wt:Z9D]6 %슋%e40VƩT襲Smf('jE×Z&+\k H#5A] KW3{ۚߑ`A+Qc?^3=b'wh~_4P(4š;(ح|χc;]i{(dYcEzw?@k$vLt9BXg_^)Rh!'ݖ*Lst ^9p^$oI"cJމS} 9P)MIxZK!{(>E+-D9$1ryG`7sS}(?ӋgjMoJWUG[?{b6;M,7>C, @@:; $$ #%uRFKK>{ᠫ}L^IY"*Wy%[`9(W ]sszuEt~&w⠓R8yp-a]@Hn8t&Ku0%hu0YgWg]tKY붇1 G7clhb׿O&eXel|C[ҕ CcG98F?RgĮn|MO 7}$*V@%E**$!d颓:Z`c|df +C:]h x|6LN .n˼a$CΤMS|Ѷ8g:\LՀЕFtr:G7cBv';AhO8@"pK$̵uƘr9C )4>~{?1X;i 7s v F V6'þ'Œ,f`R^n\.FC;"k [IA4'tв,%؏M&t֞sBŧ{>zFh:}ӟLI3XHdKҁW[ I]k\sуw܅f펹O % G'vy'[.ȂEz>'۳ѴϝV߮E %T]Cs%#))FKds,4{yL`^󊽊*ٱ:Kݫ^'z%""Q\2Oul CИ^T+-*_R0ut6:H72s+XJY!k=;V[E֢קּvZP$z[%(_҈[xR-Ȩr*Rߵ洜ݡPĮ(1ّeʆ"W[ie Oh],歿A >&A eR:F D3IYek(tv,D۝@5C"*CRK 3$C;S_#x]~+XԻbQExpy:K]hs[R9D~QJQJ miY!A")W7A29>(^GŌМ)gX99E[(wZϝ j?0dYsc(ClRgJhBPG0c7C^'e:vRʸTcw:MuWSʒnRRd V̘c`oI(YM|0tOtI}}mruUdW;aU,z&OV'ewҊ<<<2]ΰKMuig\RyZ:Ď8~$;JE[M,Z{)75. ~)bbѿ߹W3Q&6PqN/ߩ(cΐ --a lxޛ:{SgzoLMw') /hsIe% r5g>=rں4v0/>wYgԽN6ד6uئNc,[SVwr 9i/Au\BqU s;.>cwxFޕ64鿢wwf>elf۰yoD]md IAi] PgՙYdeeLx6żG1oQ9ocvHE1Єkb$h P \ fShM>N +׎y v[xڱsvpM؍'\Sd0\:6-kKč(ki`eQUFxR˩gի{?Drqa/@g)"mii10C6_O@*_סp6{>lkw?Cx7~<9'yT8\Jg׏ghUnmn_7ÿ2{;o~sbg_y?w'g5Fh!?۳1Ekq|pǁ.wFU?Ûox[-(o6 P葭F37>6չ"ڷtb(HW;;G/Yv E$=ow;z+]lwvljou0'7^}G/vk5Wކ/nD{'okhpF.?kp|R{\4vov"^#U1&]_zݨhE<ʗSݓK7?(IH Mni'\Q? bދxA}2`,e0`^yQwOfqyQ=3_~*=<8]t;V,J2+8ق7m(K7OjXᚃøĠU.~g>5=i߆1𢾛ië^ķjy։cl7=+]n{ uW>, h`Uuz)q d;mm=P͢7v_ɕᙽ als9/S7:8B|okT{oWRT޴1Xlƹw_>}cv:#0#tuBTszS@grwGXEAC}@\ŀVmP@ˆ)bJ„tF99`ә9 YguTb5G!8R6|BENvS@XTNr1A#gHUDZ#':LX?&bY֌lݬ.,;?:q,T9L|vvw[dLehoB>lh14 3(O2k\+ܢz [zqjSkxZ['6|H\l"YYYY CV.C 9=`1*G\F0fzFu[&&J`N\8+bg@i; \wz \'p {9-%XzP͉:pHcG254 a`D?:X=}VWm}wE:mj  (S)l4̼yM6bsZ-$+}\Ǎwpr\gVa癩̧Jk8ҾO+n d7/uLNFR ق s:[rMr},.zrR)"ԲҖ\BKYR?tUf;3plwP|yd/#0OXT#^T{DŒWizxFKI4*-ju^gqu^gUZμi9c/Eón*Ҝ9ꗔ.Y% >WALфTr)DYLYLTJ3y("v:L=sC&q*(7~(T^h,r@*ZM1rXɕ @:sxԍt(GNK޺S @]Ѭ:.cɬOŬ 3#Sj"LffbB9$[QN'hT =6At $TJ{Q.Bޮ\6EH[ּ@lf)Hm~fbԦY R ;%‚0Hm6f)W(E9RT擈cX̷3fJ3ݛRu .jQ :=@c-^`BiJ 2S`kY,@LZ#6ei5 jY)s詪9ԖdTR<'9LBid5(?JP o)9mLYyvBS .9왖fàB6Eƈv+,UsSLoqXHF@8qT nd8O4L0읇3Sݎ$(qO}spڝa ͼ^d ⺍95<&aKdW؆`q'.3`qYkXFq(J!DΛ'%Io),%IoR⛔&%Y\ÞT®coV+˙ A x̯rͺ(63[A)[ 3V +qEb*P ɒf4~|y^lqjݙ7pAIm:}]̎q7Qq"N 04")zњOJ<%-lj詽Yf6\lx $K X|E`r sK:5g"m|k(DxjH\Bg]8bbZe)k΁{BĘήA( ӘRNL4TI$fLi3Mb.n6@wc9xlnfn0P"D@81hVw}Ȓœ =b jy s_oU16PRR↕@tm\Oe uz97;%T" L^BMsK$hiI,(ubA''_D <Ă J,h,h L qYdĩ (qU6SFrGAԊOf25bٍ95 rg*\0.%O ݝ}[1'[r(0vƌ'4C+&cIIX'E OHH/nᓉRIӀo~4{@ly2DXb,BaFDT2- Qf)uƹm!4QSv_*־mVlKɘ뛳߮ !$ـ!n86Ri̵C'$3si՘[rJ h'v7J (1ĀZ P&Y3I 34.!x @ӠG>} 4K= ʧ{n\q4umw!*Vlv" SQh{ECc$He(Z#׊kERrn萰Ҳ^,O=֌a 08DHubK><уa=K m!35|*d@4F5X]=PnÅZqGZr SyJ09z%`J[T,lUHk11"QKLb@KK$Pb@%8tL)mUߴ8`ǃV\e oڠL9O cyQ*>{l=4)dZǝ3|Ml#O`r~)NsUͪԂWPRt" E|1:ؑh)´4Sd ?U= $Q>u l5%2b$jᄶ2*륖#M 2kƨ;mϥ=Vce(&1|։߳$擘Ob>,0{*&9&okaQCt0|XI&)= #qt hMgkSҳ_+E5zΞ6J>YKӬuhFmiBtixVc?JA0Wj9k;E'QjWm`  ?!C fʘ#Py0Χ?pK&^Q&0ny25Zy\Zd'O޵uؿ"̧]$Q/).`؏F=dȏx=-X-tc%L\,k.N}5Lzx5$dガB*}/u_QgWʇ.(WvQBlGbաZ~Ho`U@VC@v v?|;΃68E'6?F6ݿ}~GL I2)Xr;'}f|ezw ~ّCYItN?X 堩eYD6e*f"RDݴ"ٔa;;br;[3J Ԍ/zJ-U0\mi[ B>` 7 >ݽy,iսC,6.]0q}O)'RI)!\$lb@Y~]FCFا c S"쥅%ڍƛԷ ևls!^L[a{#:d5.'goۇw/_uۗo@m9?ݏ@~~xs,>KzGǟC>%,?} _^E9<#~IxoN/Ν}^DR~b]]=>PD^* >}7zz^oW[Ox`uIk̽f 6=I. M,6*Iy} qcKӤcN'c'͘"N.ߟ82'$6)3Ĕ\E< ЛɹT6`k:A"#v OKqfYk@fKp,7r[BYYkz\*eA6KzzΔ i}R q)!%fMxNk]zuiہ`9^+՚_67?gfcBEdņ)l&d1K&jtzNrWLz^KLf "SXۉ0̔'ѧ5<9` A|T,0%>ߨԱ1͜C&4h.x8183g#,RpρN4*rfQQo8EX\xBjlQ?1<$7iS싣v$36鍮zjN^PFLjf]o'w1>b%M2wgtW@Ѯ_fJ7uR!FMKc>LGJJz҄<^)#dt+&V\B@I;)]tUszMi!.Uoڂx%l/9aj2d+]T:Dj}Y){/SNV(#ވo|f?C].9ywe4T!^GԪM v(A-&i >)6zL]3ͷǷFLVxO/>NLBLU3aw>"!|5|0ʰx׸겳4Wmѡ R㹱2λ[t7nQ4]3}3|Z%N߹?`rivr @pN)܏vL0I+EMzͶ@5akY14^j-ָ 3!P>f({PP1fY^@DjIeK)"~R+y%#8kVR"wV$ ءŕwZBF$tD\Uxwڝ[43^HSw07 {γfv;8QzZ^A6E^Q>,$:d D4y^>RZG- *>UEOE3(Q<:-Զ |ި&AڨbHL ) z 1DQ&f8M2,vOw$3za]y: jWh[JoY^ G-$aυwWB`.Goq9rĿn8i[J<*Y!FkP!WίF{ή SncCQ. U`<^.e7̈́$ҭBca>?(~45p`n-`C}zg$M\0HW߼'Frk:P$Mk9}cpS$8Ƣ1RL6lEe ۺƜ.$Bw&蔸e}~VYapY.mXg[geٜ%op8?O8߰$9g. FJ-n\H`e#Y@WP 18)fمR?1K`^oބ j4n;ȼܢeNnicc~Kֈz̍D{wgZ۸З@>qNQ-nM%Rr_Re\C'M,ypf83,5̿٭i1orZx|[ږ/<*=WfkᲸ- \}SsdQ󱈅hB80}Et!9U' jciM[ G8 xK^Wbgzr*)C \Ǫ$%vƽڹV-.+6ͫW4؜?CJ[H C$rs`z$S2{H>Lt{}1~Wlrv>ybXTr/<8ɓqnἃ0ӓ%:3y_!,KdȰNN4U'ŗ1T]vpvt 3Ǜ @A¾š˿ Y7; mY~w'oFKh`# 4)wsԄh3z:FqĊ36X4?Y6}bToNN̮__M.Ņ^|WISfѧiĘy'@ gB_t];12 4Glne7ɥ.SIC!gsnv½5yy]]]7hv)Y?$.YC@h,FqƒvW݂3Q,<'!(ݛU"w48o BZ}c@,` 1bDU<%a"^<]ixFi 9ōCQ&O>Xpzr6@A^Sj4bմ6I$RP1| Ljm_3+digCQ:-AڷΑh)!2ʑ<֬'t`3x=A3yާ8}4qsYw=!W=7l!'d8T4gYՁgYPd}׻K(8R¢b:1 ueQwiӐ&: iӐ&:-V_0 $WIR.H܎+Z9 WD.\<(fF֎`ovet; /iĐ#D +2fLm kN:L m)P%9ЂA@0 Y^ǧ`42/jd@t+ۜ^K8cXg6wFa3ڡ՗goWR*I#7Η RbR UJƳLox¸%Ob;*Q5X`& LBֽ݊x-a;84j Sǟsט[k쳾tZ&ˉ+U>QHR@DRI,@ 6-)0E ۾cZJʼnG_m+=l1L #= \fෳ_܆TJpaa L{n*O{l+ZUڰFX[T:ˈTY %xz/&֛<ρgdt_G"h4r5(ʳڰFo|,bwM![s}.zqGiCfŇ}bO~W.BE&}d),?#0yKxle@b⾕h9tPup3 EF2X[DXE[QdlkWE9u.ztTH]d}Ўr>ՋA3%ɾlJ ʖFLh9ISyz`B %ƸWSQP{PomAx^֖Z%wzXH[q%^eP:hna00;@@b}bޣ%<[O좼i ˡbKhڸ`j N5XڔvuVScsմ?rKҸXзO hO{ qٜ^6N6\%=l<"l/,\0@G7z~4GG+rT~WGPRPRI"s.C32pr&&NhoRc5NZ(9:R=]qԬ2H/JW jL^yC Sk* ZUXLeTR#i8c!5) UdiLPN (慁L5JT&5nT!UG`5Jn& ?/դ.\TQَr\&Cp!\_*#]K}v"է?z2u'y^UnxePhOI)BgA4t3̀bȢ(d&OgK K&9ʘLB]8oHΔF,5?'sϒlriC@AiIՅί F~TjށG\3]aѯ}Nlwh}0?OB+ŘTayzMb NAJ#"irG,'#Rv}V|S^ȓB˨f)  "YU[gp9z ETŕ=aENUQ^eV(xB1B?⋕.,wzl. SH_Sw=^&TcSIOZTgV3_6G`u![^!CK{ܕb]j/= ϛ@ T選y":7F|jR]jw<ڃNؼ=Öh KM4KA٩r著58xcTOwk PE:3]h?wmxLalCFn3D ܍' %GQ\. EcJ7ݿu)#H)cW@R*CIιT D˗wZ^C"w#qO xP#}%(M1BȈɓB#A;YIA d$+ch> ` !.boϯaCm2B["OSD)Uj)q(dUvb&C DYK8s($@# K)Rv]uJmUY10ahfJ2VvrMo}w2:5ЀB> ؀?Q>o0{{+$ $8UmY3za!TWs!sCT1&7>igaj ?FgQ oh4b6w*˭!pق~6Clg:$kx z 8+PLZyr $RW HYSwrhZI ,˝lfԥ.!nnxʛaV3TPZR !Vfs,V%Gt ,QJ]Wcv0TU%1xi%R(YU 0tDZ.eTw޼oL9Eϡ! z+uվ~Z|tuu|qn /)p@/dPJJS:݅w{ ZT﷐1V^xBҔ6^AE=w&<5btErOWwU(}88:2P-5CS-pt c|Ucٮ>!BK%ՉdlSnKZi`tr g> 갵zEr])Yf!Ò$+1PSW )Q ypJE.ׁCT:C,pr+E r'+H^"v\\pӴf򃐼'Kdes48p 홲h9t"`o9Cךs@ΜBgey/;Y4;NnB湮S!HiؚW91-qP.`vi7>P9\l4sM>eٰG2qΨ<*?X:IJT$9wdGQ>Ԑ6SϮg9j 8 5ë] v+ҽ }\@CW 2yAH LZJBCѧ%I)eliHGsǏ`UL?BC8@ Q֐!ˍH!Sl8s2K[:fa:Gb✫D֌p! O<9  l N8H cI*GQqCbA CH[֎N֪w6YN뻬cIn2m Q V)\ܼ2Gg*E8E4MLc3)ٔ;W sp}"[@e5Ae:"kEIc޲4U^rDÙ4:ryBG[YLqc!{Ѧc2S*yn(hShm,8֤*Dx6[%2|}v /-ݸK[+UUFΐ7d}a))Cd% 1b-T0V >d˧> 8Ͽ|Gp7E1?!$^緟/.=Ls᳡]Lqxos\5לoxSu. :sn4+8Q$sF9xeH=UI$7U.q¢K\)C27+e]ÏyE4{d $o.ҦD5'f+MMmbk~Qfѩ($ȽN de%߮gǂ"iK"Ur~<"fs ) =EWB:_R膟K3B)uBzr3B[$J&!_ƒ\ svB|rj3糗sf_q󬉳-xF#G6Φټ;1yA'P'sI"~R}rȒa(ix3[+.PCfJ8d}  0>lewܱ);@2F'IJb*[YV8t#UUeF#MCR5Wn p.3y6FP Ri!{I*^U: >?vOkՔ6&]^AS?15~obb>ۈc3h E Fb I]OL]`bN+D_&w5dl|'0scKP~8zpҹA]N7tI7KNJHMUdc%5T1(SLs(.'nnm5I1"sK/+".C.1<ƺ S$AL*)Fe1D06b"GNu˪uǽ8DR{0;UmmyoQjz -G ~742x3pռ)_o*5P5.G##~+Nϻhel`bc F wƂdֻFp؀FfJ}suckon]8\$̦Hg]VW^X;^~5 ZZdFt-{$6kz%U2m/Yw"Ii$pW@ǃ`U߶5 :(u:&>Eu_2Q{ǟ PSO8+fH2e_פtbM3z]mK("ЪNv%Ve=ߧ nx˼XȽؚR ]wP^2+NG!t~mQ t2w!T|XP^Vf(ۮ1jlUg46?=+R>'#I"rkf!P`j?9|SPc.FM0mτn\" MDhE4 W"ǨB̋̏e&0$H+#h{;kl-Q!I6GE."D%)nO߷hýpPcxB'c523iG1-,)ӻO@uSJ( NpN¢,^mEs"|U7r~{ʖw1Mȗ0#B(Td8~^?̘g㲜c>hY<Գԓ];zu,}DL8PxÂyZta[u>oltٮet )=%YPn͈k ">DK‹NL21% fĭ*u8,!m=y:CНJ2LcY #$0Cu_7 wTUP^׊(vD~(BȲP񴂸PBMѤs-" *p5 n2YVˎ26\@z" ЮITt}+߈ چLkHf =+H[Jx)Cdi`oނt~t -Pup١]~h\s5_ݘyLmjY 9LpqGnD*77fD'FzJk=(ZHkSЖPŠS`AiMYށ[tFrNL^W0v0-Woxlj~[#brԳ z3E4N(ly*icIƫ_n*!TGx^ϝdѸJ{/10t;utI7"mb+%*<>ع|9C]ۙE ;]4]R%$ײk p\U=yDf840KwɵTBW,pA}-ea+_rmywr018ͲBXliӻq|v]%.C[n}h%^jc70ٍɌ4>PMxo!xVѐ~Cʰ^8DXEgn4]B5jF>xFli!dnpV" -l= o92EzT!HY44=^{?XժIP8t_YDu%*JL U$ *-1S9F/*΂ox̧;9lyljӤtwLvK-F~hf<= ѴG18U;OFdȼKC|ꩳa{rc.Yyۉg~u?4|XF7Q'(kzi(T RV6e)R^ې@Z\A_٢o?yu Y-Ž^Gy]܎.G7㻠^RNGwqB`b2qwh,lz$-_=*J-?RRP7-O oyT7g'N}mZ\Bh twBF2.Ut:8.”+7tZ%Brl]dsK\ԐkAd!7DWP״`[r뺜e[00Ӡ)EÐȦAJ7i,.bt}w{y9s{\nKɥsDJ摟[n?rmx,V}n$]0v?:Y73={\B?{ܶ G4MiP;ӓ{IӹdIONuYjp0(u7\JaJh"NK29$iGalR>wg1HQ6=& u7i0X0M$>m& F?&IϷn|E>hCZsL_N~~ܭ$Z-s|y$ F Hu tA'3ޔn%xL iY j!ja6iä 40~3'13I=3{} ͸D{2,dYД@^g̹s፵gb@ۆZMm@cEI-V~:QD &a ^jLDT̎&yj1F;ɸFنF!C%ӯm@,Z]Dn 1ukBc4oIvH( yCoxJR-J2*5 T*ӫX^fl֤ͥ+QokɈ A!xOĶ WjOW]^ws.> ZpΏN-*twX즫@#lg,]ݢ`b͉a[gNG^M ;Ky1j];3*; +۾8a\X)wgW~0o׿ \`o[lz*]_vsobo'E^ܹ뇴:Kc?:֣aX1TK :b E<ɸL". )*Է[5NecAPa]82 ī(qy;anM Vos}iĀ;.ЗnXru'`F\r'A܃`ІB3kNAau#ݚCz$zJѪJF|{$Ew2 fR['w*8ʍqQ!90;3KL2=2#Y "+ŶrQG[uXj)C f5i $S4&ydbfD>8r´L ,0Khy&U!|: x=l^籌F4ވ1x#FZ mhJo2Qf2>: 'f4at;&2Ôu8E9#!RT;}7-6J6i*]E .o.)wp`{y\̀=*{ 'a|%WΑS1_c; r3F ?ޒ3a!ȗT("Nh~:V~ͯ$H9鵎P|g^{ٵZeV*)l2} QpPEL',Ae(x[3_EDy(XKqYt u,S.DRUm-V9?VM:vh>Oe*MũY8e2<-Gdyplܖ#n=oJ +;% =R\ ׼M.Gvhp勘d$&q;INbzzni20ΐe4C&p* pg h7JF- ᆪpoj& GSw{ߦ?-86J/wN':Q6GP+ٸ\j[n:%3um:$CyE\}AcW剕w-5_t)8gzH4" Lt򾇙ÑUY6.sB Jsb2/rYm5!/70%3Px%sO&c\jѝbvJ'~Z0C D޼x9䘰Hi6Ͱt05En4 .~3,W?QAzٸ9+ >cVZ%Hƌ Cs ?gw^gC(xq!xʝ蜆8yb.ĝ3rpt^'pUNYq珩9/feS ;|k k{?n:t6pft1z9 G/SOrӬkR38.90V}[~.qhr6: |6B5O˧||32፥7>+A'h#$a*uM>F4$'Sޣ_9zvd(i8𶠾Ӳ6 Z$0J(zC0r݃W$>T|p#{l;Vr 8Y  );B,C58sf);rZPNǓqRkO쬔Y0:aiEwԪ:t(dD ?bCpf 8l& Ȯcb柧2lB)ɸ!`(5ұ5otlgi (ͼҟs6 j%-э]D[fJ[ìtjuW/ @.f[InKrS9+z rLdGYVF5y+?'q-s4Ui| t~?Jl9l3Y˳9e D^*1ɇ1̥ 4[exgfBB`>h-e_}?ݚ*sHaCOp! `bn%zo ySy u59g,1̩^"p;]mo۸+FqQ]H5@?6=`MФbqE)J؉ĶlKCK _r0Q&HtL"I@c6T $ZأHy,!6TiLb4.S2.S2.S2.ӍL{Lywz #Ҁ4RX"4A4 8eέCE8csҥp8KHr 3ݤ5:_ZS*]FF/KqCZc~oH:)+Yn+wÄec8oǵ?Wn]Oޜ|χϳu1g`\.`KݮX/̯*TfA4:Fo*ki*4RkAE[0SgE&AY̻oAn ًIrLĨ)D23h{T1n'G"1,^WҒ`iv_Dm"%4Kil։B4fIjI ZZFwbŌ5S:Th,'zQ_I$lv>2#>2#>2#>Ⱦ3sM=>2p MbhdCN=? .)gA&ټJb}xUa)~`x9f6v*Ԇ+BP13\_M5n`;`luA7ߜǧ:J4hWKfU걼XaBїyˤm>&m~31Ο=+!<5=_kqޙgLH.Kӌ8ceHI)I1mdBg<"m|pafe&${e&cR,[XUze.E^W&Vxk@ )P5qj{] fn :+lm}( }Xf{QS>.V"#5ذ-_EB.[f _kVKxd݇_*C0 X(;%B7*kE15_׿rK/T Y8^nthUӽN/c+GPW}g?ؽ#]]/^=a 0sFNAv//)\)7 (͘awlMv;0q=AK143rh,cޔoYjͦs2 {qz1DPe r#dR&w]#̇WP6'|Q=u{ۇcgC١˄\1Mki$"J1%*Q$XD j-6QHH&r䗑쥄izfz%(2 wN偟['!=:' a= |-(Y`ʑ`{8tfKxx=[g&z_ϹKSni "ʑ@tS6t7|k}z( ^,y=)vG3cZ0hrJ{+Wobri?p(jhJBgQ͙xn&Iy}q8jP1 VGT R6$IP&DT86hu\*U3K9C0'2\Xg_)cNFf[fnٽvzvcSYr8Ύùݛ'2m=֤<.ˌp_UԞnޓu.\equ}0>@Hc1qoW`%*1iD,BD ("X\9"T>v;lFؾúݼbDa[:K{I>Kg$ZNhkK89 8y݅&iqQnV:&طS# $S91mSggx37O85ZdLu8r-*CQ]CZ/ĵ ׿>u.i:˔IIH)"&IU 2 %&/_OY9>1`%xszۭRHվ4'i@;HD 014/(Kу.~[$ u4IX΋O SH:OG$I ׷#/u]]eJw[$_>JJ}v*97uUP!٢G{k*ߕ jW,ZX^9_VkK&x ױf1?MNtΒVe)*统o6Vw f1!p`uq(X# 1-Q<XQaH8af IP ]Ǯzsi_t=T~tJys(S[rԃ)͹&McOFID0C!D$ B31q0ԣa=2.8s)ьk}a-4YQ`uE2]na19 ՗@Pl jP4%7*-}冯C Z =uZh,-\^[B*.?s^1%1)Hac 46Yůe/9<xVe?8T 6k69V,xc;֤F u]tt{ Vu|~sݟ:G:6#"֞#rZIkt3,YtV>o2Z/!|&P쩤GtA-QgvbnħPe'!>%'sA[xBr+PQ0ZSH* 12I7<]6t1é*1tV5tvǀe:@tf3fssi1 l޿^U / g-R2O7O#@gw2}7x#7x#7x#:2„ u?( @ :R4GJ aP3`! R4FkN$ I N9;E,I 02l^aVoaԨa 2a ԹZvs]ʇ)"@KDH C"RR"1QX: bZ嘲Xtj(G2Q#OQt6aj l-W|M>.N]NSoc_%4؞*HD&HcJUH6*л2oZl|kj`n@L4SE'49FJL z(Ƹ/iw52 3`Y'a X?;3ռu).8u)UD`TZ/ݸ q2)4cT(Ihs]ݽSZydl@<DVas/-Ѽthq26 Q{N)kN;JK4/8p|%J$|}QW{mJ`V9sʜWΜF2ʙhpmٻ6cWX|I*Zh.3ݬÉ$](o&@;:.ӳI(-vggzBҫ.*|O;9*?mbDNE/Kk 5 3ި.@g',3ɪ 4P0 ?^.Y|:L|r UV\ j Ye.bRNA) KҒ&_e~/8~>5`AtEu51\bXTT9ה,)SsbA.xP|_MkYY8C^YRQS&'dIڢ$RdQWN> °ԃDž=\!8u_˧YAsS 1UJ2C #kDQ΋Fԣ^\4/KyeWV=}br:֢ԛJvACEC3!,~N*̽W*9 bwYӋ]a^HD_2Ws?Gf?Gm= ozexd5+mw1o^^p<8q7޾kҹWF) \߷~Q_3Kn֨w?BCs|t&?/ƓOgF'a=F#"v.x{sq!mmh!A% ^Y?i:zpnjT|qfeЏ@/p۪XWuez=N8Y{.w|7۽w= fn~ۯMkB^~\Ս;tDc7eV@hz= x> kd%@`~t-q&/g3cw309:,J1v]]U)kC0:Y*1be1j99-^2;C Dj߈KV!$1F.IZ|v%Ob6pv pVnT>z(V Wf~w<6Kp|6COShgήͩtS=n)wKStl*8ƃ?a֫ym!lwMZTm=ľM15x]Jpe-2)Q6pcu!XYi 'GDFL)DLx_֌ |8'!WJ ēēn/MTT R!W"yZY]C u5%A$؊A5g"x"eD"ϬBeYkԸ#~yn]nZYww6pq9M'+g;nI,h;/8rS-x}*0\)gi-,Ow._3w&-r &S[29K9=-:,.CYOarNt>Wrf8wل:VQ!z!X4;UK)7/xHMi-~ {20.j? _M^עd{nL /. ڗ}H%g3-!2DHhCh/-_2eهs:X;㡓'YGwqÚBb!M sGYlIdU!$tc͊Q hZRYs}b.`F .83qA"1iA|j"wT08rY^%c$f=8Ns:N]^q;i: ٘AtzU2D !"(eE;v.J:P*B6>œ{&S"Q?Psl'P]9(!(JLrrIiU QD+Y3x:#>`r]$4u Ǟu6HF19;rJV2H P!U2fLE9ѱB9"DGJ~$h!$qS^D$9NBo-znQ`8rϦdoւZ/P`'v*m6"7#%D؂_mύ@.RԨ2"0[H%n bW6tLY~6+E̟N [ ;:iYR{MQd1WNVAAHq݄X{l޶:AYtg(SŲvRbg&ښ[+sJŞTAPQպp5t 1<#%Z&R7=?:ƒ!y݇juC5aEV x.1:Vf񩆬+;N:RH-j{ C*OY yiY9r)j?CuE^h1`^T۶B묈ȢeZ<6YA5j*-to‚i>}v#%!=O͞x/ÞiV~ lo7{6Ny+Sm ȈY7سi3R -Ϗ@İgب'/R=ԭR$WÇO,§5k׿c|]PmV|bQ6u6x4QNPH8ǰ»Q䑉/|앶ɨDWvlQ +vv֗Dj+q^S^gZ>(*N;Y9E,AʍQ.#T3t(m3V ;풷lEd۲K;լ"tcm=/hr6N. A(3Bij 51b8A$<@UJsc|qaZmL.}AZ;쵽dmzr2 ٌJacj'pcBx4`Ս+dZ7Ƿq[ԪWbZ&ம úySF[nό$t}W'79~$齿O#2 2Ժҗ-qA2%vVn;$6w~%e6$&-j"_p"A5iiK*s2SE|ZOVQߗn>c'NDӜmKuĸ\YtnuL2J054̉drjE+n3qĸBN \a,qɥSEU Ge=;auqzEش:_δ#͵-T3j]KĽr Fؙ;WպtZK|gW!.F빐ZQuV=D0f ]c%oF6E߈(NF 6mKlQEÚ(m9\;ﲞ/(<5=}*lg 1mwvmaZ;({ 0 &cAD[Vؗuٴj -6?)BvS0J+th>?&H+rx/A$^3Fmt.Q(Rs?L&(Vgo º5;দQAm :w^oV@FI[:rxS׃"DM[MP!jٜ=7zƠeLiuYr|CT {|# -`Ks+`)è +Ylώ' 4åe39,r֊bA O(^.Y@堇S>J2|(%hC!7KPi؀ymKz(\Ӻ]\xLjĘ?A4bbfRҧ՛"™,߻;'<,!y>Jgwt2u(@1DѠ3- u 2+؄rH2ឃ:mCPwr9AmRɅqxiP|AzpҷYPoBh7K$AnN2 t4yeЇZk3}0臠|)~;b/ju՝un0Ȍr§2#Th:OXX.cJeL:`GuݏiBl$S*#u墱n6)|EZ5oǨέ6ጰkiUN9JdsU5SժBQV fǚ)A,٬~뭬6kfXY=9|R\,j ZjV{ ?WtlF;Zd䄥SUHOVC,#gt1B bQF^s `T+|j^&ZFi_M} Wo#H,qo/#7O:^^p#;;qL2 ȤrLmг= '*^ƊըgIlRR fI5 sQfI!ԀJ2x­Љ$doYȂ&ïX#*l.?2N)&7?*)ng%Cv2k l'&8SC.^_BSK C[%.P#ZVje.Q` %Ҕ^nN ɢޅഋMvwWVD@I Tn]'XIT{"z 8P J/Җ}vhЀ85AycwļX;іx8X=Kz8ܹF7cv­@{2:a7+.UEO|IDSr,\i"t׭.,Mn|(4w9 ?բ2\Ad #s i 2QkAO FF/ĭ:Ɯ^; >ڙYKk9'}frZ5ck;3 l?;rfvgǨlCA6?~ۿ<OaK+?{u<*^e闫V W>Be+HǸhARN$ "g ̕PVT۫w@+ױ CSs2QV$\XIrpBIseP;c(n>C#vr튐͗i5EpЩKm%Z fӹBTQaE5Ta2IIl'$Jx!y|;kAATrj0 7+/xR( Ux&E Ɋ3 é\LK"5:/2e(s7Qg9jIe0Q$W.ǭ>L 9g")$yQF pCQ^ Ƒ\@Ig` vX=Bse~w;_̬t_Qo*i^Q r[/GjT2kxj%@Jw{3}q=:,Ofrx}uw(+B#Ħ{LĘy]omk2U%.W%?X~b y1*ocm8Z"&xM*I9@YkvZ^kt Wyε p?֤AI-MkH~е#W'%6hT[54h7Cv=4('1Ma_Sۿ,+?󱯕%{?K^o_ЍB |Pw>OGCU]ߞ.$dgB禶-a<፶ wİ_/>%9v-i׎T4H3cE[wtu]p?] Ȍk7Ov1WS"_sk=3Ow֟o.zw-{\f81Y`]ۮw\N&{w%HOaIdtNYí:l'W,x$_z x9wjֹteR`JbbUYgLf5W%A6 y~?xSۜD9D63R7z°+p}Պh";4SmK ]7e\7xZMTJ2]sfݩ-*GO{`YH˧Åܳ+zU9 rDe\tY Ri[ "̩a rԱ g# ]qе.DfM98E* `aT"09;#ƨ$wnNe~I*`[zhqKܷ=J\A07AV*O3iS?>]ݝζRzsB.{ ?R(-&;NQQLTL8}C̱Уf$ǐUtY} y@$lk!PVG ]~eu7V{ܟ#Za# 9]`Բ.h(9uBDG:j(``O>]*T=qb #~tL5]e۬Xc8غEU_xg6ۋrk-A4(  !ywOHK%V6Fje%mK#C0k5xi9G!Vƒ(Ɇ1T KQι"3Ms^ `4QJ5+,zX+\ʻO`UYơHVͮ2B8+.R|4 5$}s<_],4@ؒky4 #$xW1Q-DZ{TU%)Ckܓêz`3@R^xyvt`v#C3j6,A\%e1BFiNt!6Q V .b1\(>5K}Fd/bw翙kފ5ZZaD,8r(sʅtn[m/UIh||sr^S:Mv3O.Ƭ99eRdYby^9p:KwIaWkoZڦBV6(H+|0$ԙዕOr Kʀz4zO?]8ykwF0qJ4'}V1͘LF՝.ZB@A!is,kqG dBc[A8 />]ԡTV[׬d=xU"hI>v!Ѡ*e@=m) :\ݵ}9J3X[4cg5 E ˼V."hPoEiJ,tԻ$("tSv?|, (fm&QWpi⛙Γ+5cu2Wl 7}5mԕpS ӕ6q/t ~s-]Ef&JNgl҆?b+|*+@ʁSZ1)0T)CHEa9 Ϝw xLZ#4"rUP ,*cѡ+Ǫ՚ew; PwV+.`<o5Ԝ/n7Ұ ϣ_o&#?Tq3Wŗ;Ew ԃxt㨓<5DUU| ٸ3&{Ƃ2atM^( I4_/!{@ވGa{ݍ;]s ӽEHRB!_ߖͽc'U,ZD__N\3Q~' /YKsVҜ4g472PЖJ5P11' g2k+S TFNVD߱ηI^Q.䇉R;)ngR$MV 5O/Rk_.I_ދo,nFXX5`qk&f ~zAF.6NPɲ+d$s>۪ mIT7-,!2PȺlr*d#3LtEi9ÎĬc`,;u Ɠt)sk*85*k$Q(V?=b÷zv,D!7}gab-710t7/',ܝߺe@/h/ZYՋ7yoM\zQߴ^`I6߽}kIﺘ=!Aʛ(H`xd)scY(ѡtoM>94%ldji4|T;Z ԲT\;2QvJr(4ǎS-Bb겡$\KɟΞ2Ԧzz߇XyhZ5O]RgfGlQ{QYYbI{==+muJ[ݳVV e=RAriHG$sM,Gbp튷A ZM$V'jBo$Vݏ&؇^R ̺{)|pV:%N gSYSB[ %4OC:(IX5%lğKZ .K .1O Ir[2|'PaTLJZʋ.֮q8=*?P >$|uX.H%Ul_fTx{rKLK|_HV(1|H0'q% Ƿ/ZRN >=K ;{T=\0cԀzB0"+i֩PI9]TNN`uVc[; vM{NPzct#<*f< 馏pfn_=6G7uffȀ-)BxSo5$$4zYHO@ϒ9JRfe(MNutȲJ.;:X~S Pz 5\BgS*;C8C($FGTJzu ˑ+4< !S .p}2.)!Ȉ5p1>Z,Z4Z[^8r8=}Nza\ǤI0;}Ca6 5sPEu7b pQr >( $rEuFKVe 4=Q:G im>]Z7hO훒lߔd$7dE+)Y}+(&,'՞!O!&Ɣ*z;<_>U[rー81V<Y"rǬ5/+=S$^@}WkKSL~N{7igiyٚT֐ L},)(S[ mqK'9zv)6bc)&vMv\6JXq$Gk:]V~IRQ0i)x8Bkʐfcse$}$Pś"6#h }0G.FNHAt(8 Fɦa¡ ME IZIT|=P%JE )5#^I1暪ʶNGU2mV{#{#s %R]M^ʥ`\<--u["/ ({v.{91 ACɂܓtHjt&`=]DȹQe~DX@!}|yvYd,IːfMߔon?iic:y0v]ԥ>?o-ָT-/Ծ]yy&J^8+[ߚzI~(b̅"=G!o5p=*|6bѴ%@YO=nZ(PA ]v۝"&IO[n7Z݆W΢TV\8 |Xa,|U[zNnT•*lXcQ^W9vn_U7zx IBm ~s9>ub.c*:xb炲ك'k#O_۫KvrSU"v׌3;lUX %Z^U7r W(w2Enmg7e#v\8jӿr|=!Ds9Z׋\W(6ғe;u/f^J{r mjh43d6{Qm[`l *6|[5L1EUz!y ]P@Á-iVY+*|(oEUhϳe}TU|pr7teSc M}tgU{.Dw\lyD=)˖*vwBdЕw79*AISNG9]7W}-Z, l ]0`ceGYI_QjcgbWX"U˻E=_-186)*類Lao׉I ?."/N  6ZUި˾Ȍz*"w]ӽ v+HIB v{]XJW2I.x.U{ Y̩:YJ>񚢯 jfGKf=^x.1_/x.q7\h%Ç3pr;c(@xSȵnG}Z4W¯WW}E=<`|-Bٟk"+Da0O1~w xQ.^K̹ס $!6ӽ/d15?7Ac+{~_\߬sʧ!Tl7]Lm/3$뙟mb[Oh"<0t,.SZW]SE*W!qUlP=p #%( ]JtL"2Hw ?tfȆR(C%Vat G!f$9ŕQIB|j~^#Ai:}.ʱUlB^SPdD\׿qUMM=hBNQJ\)= iPbƼ kUݘaD:6Ɂ,EY kEBԘKFh wʜl6NۛOxzm8]B|ڴKInDے.4rM>Ϛp6vAUG{.JCu>]"e:Etn2isT,M/ů|ڤ2<9-du׹ԆZ^yqD]<$ hx"v uxrDD/$l Np*O͉l~7ࡖ @u'_>ep6QE3Q% @>,E.6BCǟIÿrJ{GĆ 1Ztqo?_}IxxD3_~zޟuS(ȢFIzz&T~8w[WEyG8R /^i 5l:Y`Ck;?4e/4ZROܐT'ۑ߶SJ|1Wu˸~ pS~/AAXߗPzhFe_Bhsc_ERy;`??Ir{K w? 8A XWe>7J2CHn*I9F ;XE"Љ &Vg<͉c,3&'ޱ39-ZZXMǜը$yYq)yO/m)ڵ|m]HLHz"$%5Nc{ت)S (-=q"^>BڳSC"C&mv/0=QPXC5!悉ƭ !'*h]"lfT%4޾m>n|료sw ʅ(YƻE21BMz4)vuޤmY߄HQ'H[A_/J]Lb&;cmm,GmKOd0L{FVi kYjH^0@G'\iIء7qѧf Qz;GGwizvGxu;M_m_*8(bFx}!"y1T(ҕW.Y~ pEDJ4'*f'fMg<)qMH{Y1#N!+`|u)d=2քloэ_CRMpG h|lCcr)FWE oÛ73~N hNB [BЄ=Cj2|ˣ~<jM6U pںi?>Prz*MWy#L%'@Il~_:|a߳ q3 nb!h#ɿkr8O{DcT:ib( ڵP7x,X`fOV3TN) ְ$pysRR 2P̀`8tL( L9JZ" Uƀޢ!M (Vj\w|atsmBʎGuxch`ZQ" }IDCs03#@qd:Z1)d%)Li^IyNRЀT8(SeВHP*9i\}ќBR6[x:CZSo(hoonѧ>)n~G3g'AvyӀte5vcE&qIQeI&qF"q}.^D=~b%U^Yaz@v1):Ԝc%'_{Ecoy㐺G\w( Q43"s }5LK_fý"BMO@eƼQQX@u{dտDI-ӽOr2. Q\Gxwsq6zD')Gωco߱sNb(8'q7qN@(O{CZ hGsQzQ18:D-N1R;FJ :Qj!J[}֣C[ߡvh7O.r2rhUDՄhgW7:-V5p4bv!NVQf~?,{g,':I!Bb[.tK35W)s .TBl et+oZGϔ 7( DV-wn(3ulZXkaʘC3bb Hq*G3Ai2\DУ SAe -Q%4 FU{lP~b\59 Jh[(!0V Z6AR}mS5N*@n}m|9j!ZEky8 d 0VI۴ ReMd*kޒA5 xmP}x9碐t2o=uRԑd"yPwRgKM,df@C-ڌTba uA)\23/j BkvOz"ĈR $кkE"1ʸJO*&\K fI啬XUi&sD 9Ep3CefFy*$g9.Mher' ƟVKW:MqmPͣy˼u[=ɀ28#&hIХ86FAeS 3K XBZ%EM)>L] s!Hhib$ө۾RExԾ)4SΣ9j5Ϭ.O3b/'SϘ$+`24Sjp=HG re&JA?5ՌJ;g5z)T`={dgvŸ" 1kc(|~Cb&=0 =Ma@Ve1>皷TWWF/SWGtBar3;7}4gw=jmGPiqy*w!!YxfSA\ڥrS=,% $~/jd\cV =[Q |a3AA pԂQ@ j0pdŗ%RzҚ!rU&m` 9/4}mհl/V@vY>ͻhqUɎ(7`6tHŠP54ts   La$ND)3$Vh&ô^ťP;?E1p?u"TNJ[fV2VEu=XqͬNGѢ2FK*,+QU)u]J TDK_88o&_:V³Z,.w4ޜ/&ٲ;ʴv*YBdo!ZR̷O U~I/ˍjߔrтVI )?֥=dzLr*=&13Zj̜ΰ@ h).R`˼_e#-r#?BP.40Q2"P xGH&xk3rF%^a-i9s#EXKEPI\J$)EVT]6za3Z/D %JYx@xbEa*F߲*޸K9 ݫwz}x0z-15^ F!U QhRD֫u!_%n̵Ȟ8P|? xC42ZS[n{a狗!w:Ee!k5?8,Vq Z;_-곕@ p.WŝʿdBe74%+F#Nx5wBҚ*{z$%PR^G "; ɳ)"iȰC&C/IA_- HKi~Hy]km/RJ\V]i[2 Ap&.s~4e3.y"3 cP.8|"9Or7.iOIeg.3t#R]qaSer9ERRmI57dv{ 3 `"|$xc|5?4?ykcKkD cʐ9`RN*>,6X1Љ(w6C&x!.Dwdbzpn$VG :dSwF1Pry _h8GLrR:Dž88Sܔ{@`qfуg}dnۭ''!zV@5!E#uq/}!GuTCs%x[b |ѕ׫ꣻ_KW9^$nsY&bNdP  8Opp0C88^)1C93 A tJrw̛8g>&RcRE=}%-dT9_,JL_q% hΚN\_8>%yw%1olF:T4,1NOc8|\! Zb լ DN8a뺘50R\eDA;+^n+fRRU%rNbm . (N$@($#¥hx<~;r# - U`d!ΕJ2%TH0!T(=@RWi*>vxOB[U^R&Wm4/Kɲ'k+& vP @8b2-KII˱gec3x-oz:ڿQ(t~Kgp4uBh`LǼqPaa,sV('< VEQyD•aW@!3R_ (bx:e9J28oR<訁J<`d$2t:IXyOݭQfCsت|CsϹ1?;936IζVYڌ,@MSlYͤ2449*c~dFD +Ŭ@PٗF[1\@eQ+tp"C_aQD6iYs8F  ܴv1 g7Ql"':!_NXIOA7\Q0XWgq&љ_hysu/{^D`Č bkoGޞvx.d)%oVOnGņZ{iX:ƫ(>FӒ K-*JT_yΊJ0iDuM{~B*n_|tVV!|? E*N㠕 S'Z1뎽%0sz]˹u\DeYFIVӏ5pYl$U%uV( zVo]ߖCsY)^@UL(CHgvUhWLJ\{aCv?KZ!` -OI{w?*+@]"Xg[+1F3 0W )x^ū+;/VիB\_z,.0kKI4qO"|4\ }{t|v=ٕ!vjWFJB"} I?9b-_sRe*{\2 ~E"h!*X LcMZ@lr%BG`-+E9Amq #<~H̯aR) !PXS3nF#޶S8&gA["D(vo^z NQ̓Cxז6\[šD1 }_iB}[^" 惧 .蔾^rN_ 4)+@e[_zT/ ~{;LB|TJIS"mM8eD#('u{9Y|j-)(J;G +XtcAB[>m;㗿rT;F3hu }ao} // PGLz@3>8ii<H~4MS(P_ yuFl//¤`lb.Ƀ6]5~oJ~97K" )Q]Һm;0=嘉7i'XZ7cFY):c XGXv٥r/ xc|j>}0N{uKց W;UuJ4$5廪+=s`@RB]":;khY1ߥyCD[c5M[>RcL*ɦnj`Hl 0p:>RV`4੾R}'KIF#  <="''##8~ͧts8o4|'xi(Dvz"綂zlԔ÷ou]e~wWGg?l6K'sFz#g[U-⨨9TJFl=`8~cIR7ݥBչLYc ˖du'vZbt 0BTeL*KL,Œ0 Tɒ1Vu5䍝TK]]~XYf\ZW_}^m3\4-_,L>﷫\'7ninrOwHƠN俾rww5nor#|uևսivC>Ñ:mˋlN3c7ڳQ[5ܤrIC*SCS]<=Zr:떊Aꤾ#ƺ FXZn~hА?&L|`Xb:n#}%nY['Wj\~Q8Ѻb:n#enPևUtOTNB\RSGHYj$_g#.q䉔_@"=QݖO]'ɓ=5N\6Rn젽z)Of5ZZ(|tJY{ؐd 6:IHj\yE^~2'=ѹ-0|Ui1hIG 9 9a:M^;$Q"Rֈgr.~orApb-E3BȬhajB j:jU<_յ3 ?7rw_'8g<{ 3 96nڗїg"5rH}\`{[g(7ՇB M>&ϑ&򮋶4G%Ȏ1eO(,FNaK@,O޶ JRyIɀ5?żcŤ7ۊif!BLK׀Ι!bAsњyy1묈,$B#dDXC-~#Y/{rv{WI#N6@]RHAv~4p8a9UUW=Ub\8xjFa뮜''X%YT㉱Ql~-!-=W)Hq-|Z,p뿧:03#`O L8t4Ҁ{,֦:W8qcg/ Cթ]҇ѥnЕs/ '蹋vD1>ݭb(-(lyY0fC%T#(xF ޥQj7\ԧF0bWu۽syP-yb,t$KdD2A'4}d.]^+-tW `QAſ } sg鲟zF^6O!dqJ?S̔;32-AEFqYu&+-;.?{ g!'&DsYɡeu٣ PKuhl VD*mw' ?rق8&uh8p F WũũiUm BZ%X]5`o" tW~V_ <꾬6j>+$aB`,PAHwU eD)=gZ&YVL9w4J7ֹ"9\[1'PX!$ 0`b:zbWr=7wqp3~~} Mo{Ld@9{%e(a:. BG6FcYk)\ uKʃw 2$P on {wOm.5.LjUF?5Yúea6:-bJW(; p+sgc.2(HnA]46wAA'ȀP8əmq\J`]%&S|~i_=e"v>ܠk8w7~Q௏<:3Oy~A$SGU]PVrX&kWyU$K53i^=\aZ?|oJV}kRL1H5%+@XW8# \bdZb_iZ3.(lX9o#vY~Zf_ŵ~3_]-UJ뚹2]9Yŋy/x5_oDfϼb[ rHJY٬vjsޖ3}%hM)|eY?ُK=18JLcBmn&Gvy:u.,(h BJ$JoI2(8‰ Q :H>0cYJb)m ň^oR/sᙧ̪>me&ȩG(Oau~QʹTIj=?-Tg3'ʜkZ'yBY9g& J%IpBH<3T sٶx r+E@.@kw9@d_f*JeKs/g,'x=OQHN^fK*T25՗󐦦#/a#9pu9\]r`]zM%5=t Cڎ8~5XP)emb%aTpa?[W+uO\S"˅]WT!zfE_P*G x5:kuUL: Z+y-"$ $6T qLS1}PTLm[8N4y^ q0NLa/m-gJO 'Fײַxݿ_mG%P瘋5Xp@s=rpOT.@zu&3l%ƟҁDkfs>⸦Q ̀bQ7{4Wad\s ͋ ӌN7E vd-!&P\sB'w}Xnm+IɡOA,lxڪN;];lJipo8 yoY>$p-HD6+8"g)Tʞ*+6Lz&]gCV $΋h ņ}:C[k9\] VFjaB(e<ya7#/gca$l =Ol=+XizxXܽj7rhr4!]I9G9x\mNo!$; ?nq&z9#̎jTe¥X#gc~11LCG'Zj7|03GA)_~VZD+ޓ6r$+^A<@?yze^̱hU-m=H*%f-eȈȸ3[ѧII#n][NVGD x;.H'D ç^W'WTSS7|耍,5'|\F7x,mߒ{GдkY+t.?tԷ%{PAr $0&4[x 8_] &9^LI7x*KOC)51i0 ׷x4"kR%w?o: )FN@7J0VŵOXUjS*V@ˤIJ$1pvE@É,|1}Qڅw^ănL دkB7IaHq}{EӨu<EXT0a-_w_r>hcQb]y#mAYq!?YLڒCK#ω;FĹ:F$î>)yhUF7FÎZɇ]x6)8e^MuZWܛ( {Xd6,5˳D0%FjVnG>Q쳿N Z3r٫. VJWߝcډweޙli'۹Sk#NÉ}G1^w*>Fl=FDcp;Q^sVQpZpm~Xn|Xpiw0nۅBKpFTkѹ>g:@[XxdUWV[: OƪeҭiJh`#" <`s:b͘ U*+D U(XjYcMEMZ|Hg@DM ~/KIr$D+W;PUƂy F4A@1_HK_}'B1HAkЊ+3F ^bσw! tsD8p>ʡ0&y6ÇǨ۪5`OLJd6Miˬ)fE:]jZ`o=ŷ P.Cmݽ.. |%BREX<԰w~33vbmOxZ,^BH]L"iRHyp x,kZ1bT55V WNDfL$q56hP4%*4w?Z9YC4(hys6qpZYqd/V+R/-k1V>=jmEO&?tt@fH _i;XB XRk#캶`R{d1&M?}:‘V=~<9=$3\TGu@s6u7S |!@V IUUDtp:%lf@CDqs j%BYti(FSP`$ޡo _0L=+w&[DH#CC($KH Nb7}K}9Zg~ 2G]-8r{B!QL/)Z<|}v"/_=( Zyuի-7_^{0k@L/{5C`sk-p9J 8sy[7 SbuΣ18@C@X\->9DvcmD)Ǹݐ$}XjE`Ѹ1CYnqdglvNϋE}\3+߆_?.\n0/Av]H_[?=u hfć?>}~WB&h7__&tdm69lx1rӶK)TR|{v\qqiŞʋg7_z~?Jn.;hW\ɍlޕ7}NGZKKpb[J_ލjQ #y3X1F_i>"*Obs+{m蠢NYb&f/EMt͈ѽ )=T1$b^xXc]/R EQ쫊b<h?"$\LM|SKvs>Q4B{wPk| &}JaZn.9?~`&*+uW먤pd(WDozs0(Q2gZx:mэ,Kc;]wxDhnY83M />ow됎!kJ+26JE%\9cq >e)44X{Iki 8LuSwSC_\{n_g_^-?cY[|Wz_kqw9JRTq^Z4Fg=DcR#wXZ:*>WZ S!È)LVF8  dr*nD,ǞUa1tiKr9r`s.' >Q~lJ0pB4` M.$CruX2̥gbdV?/Ίj=og/1g=PIfsD`a8kc\EYIk+hĝ F2-$ ZaFfG>X˵f`2VX2[bZb[2QZZ:>hcdvb(~AY(%TL8o6q-jHR1I ;W+$5 QHMl WlguyΤo;9.]h]-Iys3l 3ɥ ic^x #dfg ̳L^ky/޸[!??=Y9puC6[FwQ uHϿV>t-?3sccJ ]-3Im{vH:)s5z#ItyzdI2Zj.f,-x}a+]| fwiqC4iyD[,ל-Rc_W{20[9hD~ EfsIjbaJ JKԖrbk# #2AbOF\&Z)zKIˤn0)} HO0r* "yF󎶏 za@|t'&||I CT.Io#8 %~A*HB呣H5OiLIr I!D_ R9Iꩬb2eRZ Os#2SUKiFļC3s^秆9,YLs-50H7pM'$4H!l ۅį⠳ m|qӫFFMכrnm@FMrg&Wg͘v|Rd)q Iک![%㖗*Atl@4ZIX UФqk59@12S#IҤm"#&BKAWP *{VӔChɘu-1U.}𦌑|5y3!]m֒{h Q`u'j9*z?aNړծFF:Cb{F!co7t0R̚oUX:g4nyIjVZvHAZm 3l#1ZIVug~9rFCA %H+_`<<=911TF9E 0&U, 0:!G 9P/茼opuàoܤUW|jNsRzAGE? ܚI@I?(@ONo4s'|[w1oiyKY6kh>(G1o>{%5P^rKgԞŏ9ttu/uvP˲>MGRoJc0`Nכ{x.R>>O޵57ncV8STkv35xmu["N%Y%RI$-$Ds?cNͶc{{bmڍ޿vu/`[+׎V.(*jq/;奺nm8AC(H)Е9Ͱ*✼G[skm+K$VhէGx%!Ikzntp cntG-yѤHeG;Ոev8>; u{5a yu]G 9LQp7)qt;CzqFd을ywQgy3&\dHe:8C}g oF9D_W ]kL혃-J5"A- *N P)@}EOˁ ItIǂD3PCfp'e,? ˘& 85 3vm.GZs,JzXU;RsׅwsVa%3,Yn%(..=,ro,faXjTы;'I1&LK!IvOB1ZKy~hſv,AȨ ^հfgLEkր%sз~_{˷_)d86@ԨK*m$ib]DIbRrJ(4!IJJIpXaܯe:,P"F)Pf"aN"ތz)躺.3RJ-6a\Q6R̮6d[:,N Fj^*,R$T'Xe̋F,("T$ CBWiq2V9VqM ^q_ [ӳ>PޅjV &i38I<1 F yNlh$8p:!*ta(Bę2QDZ#Aa0q,RR#f*tPILu7yu8:1r=؛mf7J&F]LRN[? e[%スjޠ=,nvVû -mue3zIzE{͕]3ǀmB9rfҽfE Zn}r{ h5Ю/K'jÆPa\F4o\ccn3xJT|5ev_dP X֡gnt>{mgTgjdDo1*\Ս0RL(Ҡ7 DTV $LIs!^DTqH[0CDѯ#PL6j[V~ԬW.MRF(0# '8)5$$c$) E/R,9ň$D?uqE0>i}FJpuqKD-+-ĭEboڱ4G2de~jehdWNMx]3ǐYn.Fo,TCf*pX,þ,ϯQvS=|ulmоی%I~iFj,>p9-;⶿ M{7:X 5peJoBr߈v;g,F'CYA9wp E+aAxx6.Xvczuhq0w` (5s_P+s$rd-!6Av\ H 1$`TT(igD 0)D@BERƔpD8UX*j3O> t^]ŕA7 2xM),ӇMlɔ?ş&xƏ4J$_tywYY ÇMXccU7oмJ25ǫYl^\0I I]x/QPO~H>]Ya:Pa66b/Z})a$\Bds6Dӎ J%$@ԹS'X(t% P0 w=-=-""u, /O :}8Q@ S;QvUŜ VwUm) HVi*u,o) ܊isފ*5qED!2)OWv: {ީ>e<-iK -fN4nUІaNEкAKW/iWY.lW=X=LjS֫Ńvuym(.g0A>Qj Ň uԂ3wkKArW P {}i{i5*w營%t ]Xxeoz5*BVO'?mv1'ZkosK"FK3jDܔ(!D u@(%'K6Smi< T0+hN W v1K!M3687Z(8ş|ee^8#3$^^mS|+m8Zh~Yf1zhLG+KYoQyّa%t^Kw{ P~1G#\`0&^B_Z 6&&_=oB$^)Ւjz*Jا}_e'KKÎ.4C>iBEoK:i)"3G]Wt.HT]'0Lʤnd ®=6,EdRI79r*k#{" d=RڲWQկ7ΫٮX@:{ ֊<靽vj}z_OתAղ#*$T[ Kr ) ̲0=E@Q(z\,0> v!(% &z:ꛩ7}P#+kOw9W5A#>ֈ[{WKZ2<ք`J^>b~PNo{=+R6㵨9N՚ԛlC޽`'oy{>|GMHG2RL%iA>?tR{I'&lgqxm$^%3} 5L\-(#@+ɇxbޅhJ{:\2?*džLQՂL'a7hdyV41zFEs4FUy g9kd{ /aMShj?P#ihl $E7ZcStTv\1A),JUvB5( s.#TˏJVMt{)CȋrBnvaYf.dA%4[𒛰*@MQ:8sxuο7y罎TB!K+5iO $(}J|Qn3L3L3q7(T",#,JDJ< N9:jxXfJ\[ 8A[+Xr.\aTHNK{T ̓"_ $@Z=_14,5 )N#P4EQ˘&!}+ %`l&Ʊ@盡dbϪŪA3++,j#4H1Iɠ4;N{aįdr=_H')~v .|2>׫sn"mc6Hupw .~SZKcoAfդs8Áx.3l! Hy• 8"2E14c !1g߀(6ST<Ϡb&re\K8*p )7JcUXŇZ84`ErWVv}rWByqv,/֖wax։`~&:2-&p]uH,a(dޤ2Nf_7Zono>}-M׷hZl25[ʴlnjy>ۻxu8Kb}|wBQhA2-bj'8$+Oe1C)˄I Qcď[`D)<WhXRx9Wqej0ྻ+T1.t&F^ɘkWAsZ?mԈ_'yL * H8 E8MB6 ,pR :$ * ahVh\DƄ@MfUMe3+d9V56 \Rl&bxq,+=xaq(⋹j,ݠ+si]*LM'rC㹳Zl9(߆5$UHٱk~K1 uq|o=J#cޑ-\*YCu}Fढ़%$K79R+r/)jH0 !UW@deۭaN&b?—L'Sxv&]MD_9Q5f|ғw91|\Q $dP?G :Ϥ:{Zݦ=>Fic0OYQm*ۂp 3YpmF,CcN 4^z}Oa4"mxz,Ij_ &AAwJAj=`6FD0ű`o 843tM4zɕH̵߈T<{,zջaȧBXC,ӸAzS3bIN`^(ߦ3{ӏ7~ U9SffqyGk3~IǜR,fv2WiU߼k/QwuÛWgYHy 88u$BuHG~3\a^΢ɛ1,(cKh@fQhh˖u`m`𳿶F[aL1{GEGEGEGE`t7LJnLcFc!saKY,3}rJa\& J⹱JI?\0)ozʏ~^A1׷gQ Z<( ?>{<+3=^Use& Gh'Wo$F$E/ݻ>[:袯eXqfmMq\qS Ṵ0wo)~=Vࡁ hbS`h?G: x={S\!#6r7P,(d*q0kB‹g |=X@bػ9lbȀEٞG5<>|1eEyue-&Ei e'`+uVc1%ӕc6Η j DPse V4 #9D4'\JL f63t" hʃ$LbwxpHtր ,fخLN[4 V8RD v)\nIc-1Q;aRGRw5;d<L!MpDK"n+Wy"-EZ0KF~[@rRAm!n>HqAR|PWI8|0] b*<@%QRA- ;4,ZM:2MD$T4}P#Rۏ4>3錰A:#lP=#:>D#0o$\E K(Xm4b+ <w7}P[]%jq_^-WH &/\0  Dxb 6#;4Z*#60K+^ 1dW6MС+ h|?#lO:яޜEnPQ/VpXR#qFYA3SXjfCѳs8}}9jDoQ n?z0YLnUYF0! HOЅ\x! Exdy6&)#/晔1#R=T R U RܠZ=7ǃE^D #`JR4Kyp5Lx)P@CT~ლ>|a|q,He 3Ȳ̀)< ҂[J/dE,siϕp'!$%fPA>m4J7/"ߜ Лi Z0w3FZoz@&Xub-Y }kqА Opy}F{{nigpϰkᇛ4un6ͣ:O zj>sy CK077G$ ;ॳ_7i8sClp{?F,={Wv}f~5Wwpʓѵ` kuꝫr4JЁM{+V\ḳ-z@zup~ tzʎu9|?޹%q:`2)=^x '?w0NA{nFэ/qr fͺW#8(1{ 'u n~HY0h|}e6d.3M {@f \iIoG'D"Zm7gYzY}z79?HfI9i8jAG5?]s)(8#dv^Km$A`Ca1cmYd9\| kUyGT:ޑO!*31tѲz)Hh0r,t|1"|sm"튩2T0"ĩ Y0JM1qӅsR0OHd8(*"Y5B#v}Hi@z$dJӑ0X80+4&k,\6"+ZXyX$ S l)=r0`R:"pa%˾Rae1o}딏:w)"=%/&X&C3? ˚ONC ͫomu `Y* JnNy z)[S郩Bܘm6tuVGzӢx5նM*'k.edr3ks@h0 Җip=nj]q:15tv_\HS)TUx!+@Iq"֣U WT$A_d'[o*;{5tjR0VcW{?|wR~=@lٻ( Miƺ...+Z;6uvCBqY.u;DL&aZ:֩Yj#rCVU+.Ckdf DlF*% ˍOx3}43n?BXƪI M*\]干 Tw^ʵaX[t}B^N!uu OTϗ(֜? m%±VQ;QݫGmյz.k5MJh"HcZZ;Wj1ZEԲw^r&X5ٽ.X7X2+ʨdC~*۩kP 8>W &r?eM5/ A޼SGƯmӷWhNw PAQ-҅_%h}gv󦤏Hy6ftB]`N3-L;/-z̯ᗿ|?g?}.z;N &@z ooS~f,T;Vܿjs;u3pMt.QV˼$XSpzbVSo>zRYv%&5h܌VGo )ҳ68s_t췌G.Wn&CXp{f*snɓߌT񧜭 M _a6W* -0=DP"hMOWި:p& ]s_^\nlJ)=,E7R "ps^lFu[2`%Wf-%yP+wP^YE5E 3'}a4a1QT,mRsaG33@;T֛_Y:TezUﴩ׷:?M2KOkqbݩ`wM^~q|LN;[wfifEڽzYXI?f]ȝLnaǪΑ@@Z?MTV#|˵c$ugHV]56 C#\ĮtƅJg'*ʬM',.JՎ.'uB˽QzW\꣓ Y\TVX ],E{x!+lH 2%҄D0I,S V!cx e}<%,y>K;;UaOlP ̆[ҧoo~XT|WbH&ROihO[0]rxTB$ˋNܞ'^pdRFV,U`C"*a02J0+؄HX)Lı\j#܄Jxa#e!Ud91%u1_t @t o~4r|(z~YaC=UpE [wJLٻE#fɈygSX9 oFT n8wM5-08ۨ%piA!Aܤv%yڌs 17E\+E,Q2A zbhlLSiB#c IF`T6p=>2TAyƐ@2N)qߛxXiN4_P<2T1}(9#Iu{`^9HhS _MKҐ]^lq5S+S|s*1DWahcDH'(l5\(lhh^O/m4iݤʟk Ra^_-o!֖_iMƵ*Ū䴯d_>F_EChhE8$j(QB]mҺv'hU8hp೥^82'5?D8pj<"nfLd<͒^xoxF!!/WL=MĔ'p({m. G ]+ߕ[7cF.e@G*U36Xj7|e²W㍥;YiQSN[WYsEB{bv+ݽ\aGu[9c㜕oS|xssbYhm-4ֵ3q=Ghr~Z x ZEcّJ+.}xEy"jH烫%A #cơM8 )QB`VYE VR[;-?FLzM!1($ C(D&8VHQ@x0`S9 @XԈ;g{n]tᙺ\9Mqy56R$JKZ[Ѧ2  Hnv])8׬GW{J}=P!$ǂݵ-EsW־[RdlBv#DW`{Q*@pQV'A7hބo~QcOi.veSb,Φt~n5h/!(0-vN%L0r(o4}XCC͋Fm!WԸwtX(* P|>*eth-:_eaƚʋX+!Y?+Q%0`IWV"ѡa =2i:D.Qg_}~Ԯ8w Ūq-ߟl.f͒Y76G[0 a,^ܪJx!&QtID`8 Y צK`"qSKyqa%qbv:]6XA\PBۋ5b_ΪGɐ5;ʉ|U)qD!q1rT(U[9^D#| fcnv^&@( t{xi{ׯ*wf.KZ+E<+UY 7 9EXZQֽvjDMǠbܘʚ*#ܠZ4l#S )ob'~Wji Q>AE)R-0mH"^bH5dAk%P$킝"uFO..0+/T*dITp?T8Rt{k$E޲ݝLTO5Ûi?b8JF to| Noyt/47wܛl$UM=RqU4 2:%ɡHBL3kam>e3B<S␢un52Ȥ2sz]=`FpFs9խ0O򑤕do0/l+9>gۋOUZCxqfeBMy(¥Mox*GN~U<;I߈:U쭗[Ю70@&Tp^{҈ կWq?o纶ss[kk(Rh0^N(Q?Ǧ=iLsmdRڙ#0)O[\,W18$aqTRp鲏3߾zH\Р]峳pg΄άϺNc8{2tB+MJBllhqtTIõE 7&ql(b(?t!gT/"ltmWDQ1eJ{e+@+`G l^۶l^d#˳u- ÿjÈp Wlr'gܘ2 Ɂ=!`bIQ؅;!ՉnF#h.<; /#X6_ emz2ߴo00 aNs) Ni> sPE_&r97ZK'iafydb/gQ`4+~Nu g#ֿ޼zuo|xlu|x瓯?K=}O/_|rW~xO߼\~%̾tG(O^uo#5=ӽvNܹḺ5g4yD{;э=saŌ/mwi;>fC9O}mzga[;V÷7ҽS:wud23^_d~̜_Jq ;\ŝ..L'2 :'CP2-N&6&GϞK9!>o.${2UY׍w^& TZGUweGnu~X:qtӅe oϠם/oy;$ o:zInΏ0D''~ ΏCf S?N^ouidz ȳ[ nQƝtk'z |rt=u~ L$7 M?|u &l&<}67iND.[;pn1v!7gWMejxg |]4_?8 #+c'FPD$H"Em9gib͜,/q`"fGE=pAV: E.S@ax1glSk$},AZഞ ?B~Bb\C}5Ty0&J+}6^&I(s_Ap=s"<ݬ<Ǣj+D_{ꂎGm) 2K\74=괘skBInze'&Yđ(`!偊"$F{#9,%IxĈXl0cSfb-0T$6 1XJm"BQ @o]xgbLgۛhx>_ˈ-٢e>%[4~n^K0W֌ BDrXm #yo7 \ܤ|vm Q+ݯٟڊOJlxl6e&\ 4fgRM k(M;AtK?]5뀽ގrMO-?՗oSիF28y0p_e:`F($ 12%:9 @@uhy{V$D1%_ߏN% {W۶}bʳ/!7IAޠh aD,4 3ZMmݨ(bϜ̡rh|?Pq2ܜM'6Zx&  FEX,8Jϯ@ ҫջ{Rzr{9%o$VᷓaH4w{nh"a;" |8\l4(PFRQ#o+gH=W̸6ݞt:һ;/dWAyLƹlGLBT΂sو>5`#ݷ'[hYtC_ba*@ORΜ!˾M#n*Rr&-(=9VYps *Щ's_փ]|w?ՀdI kSm9ÏYy#2ث4TX ^,V}qYo@񲕞.e&x? 9uXM})he*ΣO(AIDNP۞YK2oWEqdRp! 㶇̦ $+$% ֒Ϧ%+fO D$!z2 iGԍUZkWmoQ #/91+{C܎F#ŠyƴgB,d[E] Ϙ jc i3d e6G{-ߩzh Ӆ`YR%or,6/>8'n돎87p^}r8gD8Vw-47ȁ!zsgp^0 +]j`H-.qC x/n% "w>xpi>`١sԍ~_u ]Ud?NVJ j8z4:L5B{cq4|oG|5G{9^p75TG+slu׳]")>ȽUiA1|ARPc.Ⱦ8Vl1`Ffw`!LO <0=X΢lpHOƋ}[7ȸP>ͺ/,7:;EFWlM͋&x9;$UCmQR<'֢dݕѬ072OȆ1CQ&ug u,N4^(8}u6!ٌDXZJ[AQ4K5eN%%rk%(sd[YŽ} )ygkn@6%}{.hDV'pf $þi;ܵ2g` ": р3#4RN&*[JFļJFRg$;3IKLms/RYpeŃ^RK7Ls,$DjكT}]-<3u(Qޝ|CɇZBtIMBF ^BOčm@ٍ%د5HhDFW*osU۞Fm;TgҁOdlx۩|X,Mesʊo iiJ1a|O1?ׯ^|⴨]ygA:^c'?}tg?2_y˫> ⣼32qTV&wћ,~{IW?EzqKzNPT~~.F&n\iƟ;e8-ڥ:e3hwViקMÑyWb_9 yc]c܍{I0lv>/蜘̈́}4@XϞU]B`n4LY5bnk[SF!K|8wuo;%y_ч4)RQ8+ܬ6\a`9__~bT>?[N];v$_Og+Ƌ%Ԥw^ ^EoP4t՟I(~iꂙEG?)CW0Lw)RAd;ep\x݉9Y_2EN^ BYNcY\s_!eUEP׋\$v@y|mz2㿒Ee# sIL.d"01dDf#wbA`$S2uZHaA1!=| ˤA*ee3; Ұ0΋KxF+1b-洕FmNI!9%*G { R.,o֘A6N1.D= aG<-G Ԟ8fA@)8/$N9eWw6Q߸bNj\Oa9ngD1}f4r1Z0d a1a(J.HQTH˨v4EY\@9X hŰtM6$<{y`B_ɸt́Mlt>/R4nQ}CB`n~q[+<#Ϟ0UЌ5$ed^.=:! l:R!O~~ @oOʒR}0uжRbXݺxO)op!F@𠗿?V^dQYJSX@_i`+)Q:/l?4F'>ש^wM"Bf+Dxн#ƛvǘp^ clJM_}vWgRWw>п#6J賔PWchET,ɝ)19lĠAfA.H `WE#!D5 nNB'Ta.cV쓄0xoUy^s$ba7H@F`&C1VX-2L$أ؄s"%w9kƶY]~_Ⱥz&Y4 g0A(VDCzd=B LAPVJ+E5Hq=U МЧp"u[04Q#upc%_"mMCކD%iex91Xgl$B|7jFAFPzZſM:-[_.msFv*(B\U&MH즃.n:[=:!I3q'B-yHC-j xT{BMH7}~<6d>=`eﷄjwZ'Miʷ4e I-=Ni8h,}Q];n޲"Q JK 3K | _a,8_tɠPuU~eL:)`iN6 ֬ 2l |_;.u?LaODm[_glB؞ f}D,޴q }TjE4[\ã}kA++_QA WQdpcyGwߟͭeُ4^N@l]lד1'= r|;}\ _i6jœF9ӿ['\ctnQqʨRZlǕu^?򚓜}bWT|uo-Y8 І wv79Nȼ/qI[GSڥ5@[ * ; =r%{rBBGLE޽ RGlۓ5}-tn}Tzvi]qw+FiKPt4Jriy$]]7˧cL5SrE7|oDpI^Q@ Zn}4(Z 8Zho- (甇Szg76Uח1Q!0Z3]He"r { b_Je2XoӀ8F }[{ٷd=Drio-})F}~}iT>TKgXyV12+ӛ]abdj0%FFjʽZp<uty4wC(^Bx:߷_?\OOxxgiHRXؿ]sak.:5}ԄfZqn4sɹE:y SGN o]Gׂ#@Y]XIZH`8Zmsf# J=1O ϞC~M^ҏ>Ұ0쭕EͶVi" 7#olF$!}x!̽~c1x8a[#^ j yi%=4n҈mlfDVDvdyvl%"O%K0̞W돫0I6zxM;۾IνZܥcr/c1ŢF{q:_-mܻDZ&~7Uy#OyZɏ˵]qNLB4!jdDZAf%T#㕫 cU4Lۃ8nf/}+R*1`-=6K{y5t{ȬNpyHM:@N48_ۊҎmɋ*{pфOCZemHrqf2%%սƌKetz`P3;bY J3?"0N(I,!%> ~楣Li(8U, iJ1IʄQ1ƅ)=U-`d# K߼5,I 5P3FUwTUheZ>UP" 0H4 #IHeAIȓP`M9h2m8Z6(H|)"SNoOx%;YdR]c}Vl3^Wk:GGI]>p,ids p*>hT)Y+IC?Z֜S=Z(tAA4ըyI9G$?[p4|0,MƁFHE_! )+/Xd[nZuYԨ$wJ8 .G.Gw5ԗR1@U P,āFG")Jky=o{ٖq# xȊcArC[ m$5H6beX1qp]& e3-4Im[ș6LndgEk߸R)CKܜbZсg 0wSK`/ɸu'20a#hWԂEZO@c@+LsT3%xSB_Qw6$P}Kj˴!u|%ݜ [lxM\";#RJyeyOH.#%hi({@/T8#!}gxAeVo=~ϧ=G3{h @灋&("B_"x/:lk~~JdBY],Y$RN5V<;0#pB&tXL A"&(@1 QOJ2wN/gyNKr9[QuM%y)7swM;C ɋe>c`;P'j(0?3Z~jmoN/İ⇫ zJo€.aOI>Y&Wuft<)B[Id|TlHQ d)33<5_1v_t|ohv_J2NzѾb1_/hZ}ZDhI{S`Zc4$f 21iBis@NmڻȲlΌzIݷBߍ>{,O.e? CN]j =o煋qC[LjJo58L7e䢯]+ $ԚuZ2tq-/vsB͢*pI} #dfurHn> Ey<WyJ+ @_fK!.gr n諩Ahz.xŸ/$)s!&< 6opҘ'@Zk5X"ᖱH1f˛'mCdsNsu Wz9^" E8#57XZ9P"y_@ <6{zXn2 >qXׄA+g`61}{M!‹lPw#Zޔ#!4l{59n]\}K9i]Yv] X7v\NK0X$7{3ȗjTBٽ{֓{ ܐm&gձĎ쨗@uyCs? (̲0@] (>P6Fm7#Zamϛ9>SB7]>Ui홏'ߜ`86z6|o6ªŦd:3cCsܤ> B=iAyA3" |!!]\8ڞ &1Ғ4WuNW^3)A ?5AM]NziCz~5 9JJyO=tCu a1F-I͞&d-1wB҈"z D,xyy0] :|3.o,A?6 7aW2r;r2c*r[f=oq\;8}ӗAH~SKLlɛ[e6ߏ׼~| 6(eڀzڞjIۜlW F7~櫳N7hB}[guEYפQॕ.2aEH!GtvICLӁ~rT`N,vhN(ڏ{Txn2csg0,q6tnv9!޺k-GJT~"m\36G֗oP*țuhFBIڕwcpEi1.j2N|)*p:S8m^(/^=|[4L8YqAMЗ(X߃}ϊ$jyw0\W*$VF8sy@hA*K.2CRt.f?|8|d0iΧdvT:ȼe2}fgrju@~[?2Hw3][dEb &{$9 d"#oWt4tpvZYIT;D șli &'R sH3h.-N[>Rm覵w*w ht͂(R@nuCt/UpUdl'jh 05X 0W)1XTƊGd[1b1neN6G2JH9GNf; 8f&88rtAX?sm3SU Tu\I`W[>XQ$#Kdhѩr)3A2c!&12fvҜ3g:q1%7d<cI'9'>efH#IqmӤTnߟ&#q8r_*Jd欥90sd{huz{>_%_s"f;'+g I됄,02C- ڻ⦱sAo.y*+ MSLZPmDQ `)+r~)r~\n(2D뢡$9X L߃.4-c..fƆZ(kC)]&n Fq&"+rtgHW :p}AIDL(4q=ha(, M΢ dr tԡC)@Gvɧn9۠-ۈ:.O$Yi;33a`|V2dtp\ɵl l2(:d eV0OIE;`;=٤Qh0n tV8i]C+,ۂK\rz'{).mu Iٺ>#7)[-Pt( r٪^׼y~P2kcDمh0O#-*_;w$I:ObM)ͪA>Xλ6{leJt5})滏?~JE~kD:?~N_/O;g-%j<P {5挂L-5VŖW,J8;}^'8kBxWkVVLJ*p Qx*Q2m[[0JaMYL cVOf4ٚM}C(j‘/ڒB3y%Mx\ZE a­#biRYz2T{Wp\6q[ '~jGk7S4M [<$ۻ´㊓h7gVm%߃ Ľ{?.3[2-ۛj iIjO ̎{얋kNAK\HİW-%ZMc†dc8fJW>>Inu ІopH'5&ziIkx5lʔ))o9Vr! BY.߯Z/kA.#{e$SNVNser%8cEuǼDj5wC5aC<@| cל̭ ȕ SKLD!L*qkg̊i){lF%ih=vǖ҂ۊipKUz JN\ 6&i7H80R 1x򁓤(Yg$vY!,0ZCA2֑pL H!9]c$s6Iik[U;ׁQ[BrLc9'"uN-ꌅROћ#M$X3 72:$QŲNޕ8r#"ˎUh`t7ְax^0$KVR607(Ց2%PyUY/2<LM]C tiyl*Ђ[O2K k*sD{Y ͽO3XTВsK ѓ!P[8c sPD*JynCn&gȁ^`'fWl Hh$τ  k_$5b !Q _hkR¥ :nbh ‡Fh N hxm|B,90<:Zi1 rCsE{hS^"@P AaV[k7 o`CNnS+#&/=B:DL`6X!\ƚBShUr{P8¾W–k)t㉴ǿAzDAQE"EDtony?4K¸cb%\]5Jt @$OLj2g7.BIyW?}XU̵UVØ36535H/7ۻ/e%9[DOhwtJ7kMis1obկMV#|;?OXq*{WXqH\bR* Չ &Ṽ7au8Ols9,։ .`W_PZ?ݽAĨfV;ݷ|oIЉn<%ooC"FQVP9W<:AW0I7*nܵ|npm p5&m'ʫxtpU0 cָtXZ)Zz4[/O#Jh(WLmzvho'%i6ԱC5VBJ T ( XKFE8o2p9%\"Ѓah94Yr81D'YAs {[ͨ NPX.D ,NzR1O!3sxhm6JCM%[IE 䫲@zM#iBEHŴԸmU !$hVӤK8V jp^Ky 1KvPkJ9ŏ\p)D^E}Qȁԑ9tfG[8? UKvfDX( Rj~zRxˑhF*B0)hL.'EVZ2&hBS%׸%(*}aO!r 3"*Ŝ)gqyR"` $uJ_|$:4XjX[q,.Z "'  V\eHG?0" FJ(1'{3ppfb JWAf0 `uh>Sh,zAr)eZDռzUQ@4Nt-;-@l\ PE5v(/F4(/+ %' [q)X"EJ4kЭSA WEFF"MB>ǠȱfVHYqVg~F̉2@lT\`C8d̉<đU3h&ngNdXdBW8IϜGtf$ڠA86 ƴ5.-tfKʥ&.`x% E f$sF.IT.[i`xa.M04< oŠdUs5mTcV41<] Lś LU`0VP<~GQl. O idx & T:CJ47Gx O ƨm\^= u!] Xid. qAa{ ѻ*~DBpAؼ𖁏[T;VV9*`y5:[ѧ7}( ;PFЁ+["QԂ0|.%Y h'₤!.0ѠWZ%6Z8#)WK!n]I\W7hDp 3ぜ&JN1W|O\d |[k2R<+^&' c$/ _ ^/o^X~ŗ+fb"z-ЇՁ^m֦_c=I_#p  2[ (ӽTP UD^ufHQq&1] ,eBcfbʍ.˝Ǚ;EZ0VjiEPE%dĴvZU&$`jF`xs\ӂ c"EN, Hf%%#VJ|!ÏϡΑp-,)R`^K.r n/ӻ,@D م Bm9Z&܅L3AB̈`laWA/Pbt&[cBC˘p62.`͙A밫bJ6?ר `$\m:3jYR u`5POl5Ilo[ nZO"'Tsg*펥jZ,F֕OjlVZwnSXskFPT^,kŽx@Ah4%Θ_o6j7?'L\|>p{wUFEQ^yDZ~7.EFq AHVb8݃Ų=ݝql3qћ]}ݤeCb!44,!d"à8╇| kQ,ueLy&0ꢊ.݇lYzاps_맳4EmBT5@REL0X9q_Z@vO0!/3Ҍkڸ=rðq:k\8@ڱs {Ap1絺N$3i'fyjzD"&"p7L}9-)< J°B0'RzM #hL!@aWlo) yx`](A#Y8i3` #ZЯp4bwƵ_ŏzq-A@'"V-eq~I?,6|8&SOyX0_Sd|fWO˔yQy=Tg_b~7 wbO$rs gq1êI_5eyn>j%y Gar2#>~D$a]BΥ{B(>!Ozxw3Є@GأW%ttýVY-FW 4 MxprM*|*F^3}@'RH)Lo _|cm 0Aw.M':ȭ6Ӕqɬo& oMF rxʾv@k圓ٕw00I 0IhR4 U23AAtؾ=K)Cn<4P%s'ETJ)'nU.|tC'%0`C%I| 7$8E\Y0TN.s-i11a= y]%`)<-9#@1(|vǩE'E䬻M_%2}qks06/ϑ{ =8ܼaK޻|̈́:oTVknsnF8 tsŞ>k^5P~B+(|+JԘ p2]JĚ`Rog~Nj7D2:Uȃ&u4Jޞxt`"z^z3ݵ%}am-e//b) zqzl%^cKL]=A-3OCц 6VekD\< bT4 FT*@ 4=:g)[a3aJρ>B=ա*HRI]Utd(Qd#gq%rңT2mVf6T':?)'MݐMqx49cHa:(-m*J ?gt\On S&A;N~J0:r|X6t6%VC~>0aA B!=ԐLon?ӗ$-ԼcPLF\7҉(_'mGw˯'O2.`4807N`asϹqL9 gaetF=oқ\k[,>sΨg$8P~{M9~%?B6 g)>FBR(y9L#]Jz;l4v_=FqU%!g7h'ݵ=/8]#=Heʱ+~|{F]p9/m \ rwXAer{`Z48Y\}jJ=GM/^ty~Vmxpm ӹwaͳm,3v?->At>O_ |r^ɍMz'Ee- U&_[hQaQZ)C{eqހ~DhOs%Iӆ$Ãc6K^8b\Fi"*HPr<*%SG>oW?x~~*J_ȕQz+heΈw @k" + xhxP^.5 _7w>ɻqWw03Lt$(x㶽;F[ܶ|dVg0$STbUEQTKHv:f.ؽNb [&W(f4\T K?3)n6{z6Z1y ])(ϱ&dza{}]If<\<|,SRqu}K*\7v0ݸnRtMM&˾^Ịdjn0˽c ,x,~u"ktM\?<:q`!˒kA%%URJA 9AaNM ϡ(ALO|jÓ%ۿT؂Mo360@U*}}B$eI.Dٛ;OOݻ7?"fu6#у;=Q[%x!ygͼY%*4ݥwɗ~zق_Ќ0J,fz}$O}\Ґ/;q:Ht# ?- ĒDaOkzI{چX 1AQdۧ+6uiK8TdV6[ "q)7An'2NnLF-9,}8`nQpϡ \AXƥ.OKi8=UJ(ŷo93#kA!#-шz9! $91rZ&R`5R;PfWKz>ҳ;OT²_ v+8%i GZ2щs4qԪQiw3Ud0qs {f0Y QQ:1 `<g:HHjAĄu2XLc_0މ "I @}1MQ8FcuBZ. XGʘr:TĔ4h[s QUmrɣw/f6_;eH=rAJN/Hi#Ҕ&¼HsTŖw}yk eT~=:qquP\j|Љ&>|u~=^N(8vp)ղP -im*k3$߱ lTyxUjD+Fi9kN,7>SaBvWEګJwn\&B]RN~ kY0[EO`hh,RsԳҜԗi[LJ2a}i싛OGƺsonC0LfqOKmeI45f< \UB3=b+crR>p0DWбwR1!ZG݇xf2O(YqlIsI(-Yu5֮clLGݙx,3p`Á%*M đ bb) Ze ݊@KR7PfD|kuJ]E &DEKVBZT޷i *<T=(@@@0 aW30D?0"&2t`D& E&(pZp('@[HjHCݾo+=g~`~?=+&?l\a.aL! u&YfO>~,2 /ZTũy^7*\$ʾy3˼**A?=)2"hoKNuNBՎw|Kmn߱ $.YgdEg:P+@vڮ M(g׷71囵URm')p?|3suJT(TB_7kfKfEcK"X£1PXL2~J cyl!=5Z[C޳UU՞k 7dee+D#V eRU5w>Oݦ/S>; 4p߈ ހ[Yꃯ$p+R7-.aQk~a/w TxYXX8n F~0["s jT%=X0 `Lː ~tdR縚J-`@y;.bTw`Fˀi'7ViU=(ʃBVԵXX>J \>句A;^Z紙hF*'sBmp;e0hOj#2ŠjlOriAfhMngwiۗ/^mDZ3XJBµ$N:\6WRVm.bfs޷}/ȱ8f7F7[!!%s%kNn+MV=T##5Ƿg!Fg2p^8?4*Iҍ3p7U%*QR$jRm֓$ \6ȯ67]OWoH.a٭Z1tٜJCՊ Qgsjtƨ6pD;Hi[Ge e1dVPuOK.S/sx)$kkF'*Zͳ`w\4ϛ|L {le}!q]\fm3JrJ] ',%=wkGU6kɢ|n.i%TEgm3Bb`VCisѹbqܪbobQ[LqkYTOŭէO-$b~Pۘ EQy0Uin2Q)4R Mܛcىb QP89rϫsmG0FM;2N`n1ŭ C4ɻS`qN.1Y1o~BVDTKGNz-*_LT$Nd>&W 0^sr/6OƮ'uUdk˦T% BPQҀV'3]O%U')Cuvr7 9{D|^ N 0cΟI %Yb 4c HdXp&?F GD>.b2]ȸȟ/ LP$)@2|W#5QL+GS-Քa8JQ_q$ Pw^,iIji*pR-wUQc>ȡؕMpA!;g f}1ZLV%G׼DzPR>({LTR5tjTwTEd) テȺx.^lDi7\ڙa>g#j3d-]mу=6f:^1كJ|/vujh<>yG(WcߕOC634gU2\Х!HIv4ۂ-P RT,pL.*ɉH['qrUzS[ *〝JAe Ɠ0SJqHSʞ@S&,(g;8bXAk&2IJ~ߒU]SN7hR9rMdF^x*57MQ:7ݦl?9XBy棕1||zg*Ӹ߾eDug(moE9hټ r8:6^Ֆsuz4j-9`u]TRB3# MNE}4cRN)6͡W9 < K™R =;f l2W.骊C[Jh>{]C!(#"IOY#z<Fޮw@e+)ONzTV  Z u{ck_B8mbraH ɚCXV\HLn-̒79r!~(}kWQ\l95ڋ?mPo.2 sqa7Oj@G\3.iU'*i]kTׁ%:'Vz&o-Cf*L&IT2$YBIT́s>"}匹 9VƂ1NP/)p$TZnϘ*N!c: wX[c5ZjB0DA(qʴ #Qم?zJ#L9E"4:R`kpMp9UҘ?`!0،39 = Zdb+Z>Fp,K!5Z e]J㇏Y#Q Jp]n1))c4Y"`ؿtޕq$">aW @3O6<-[l)^odwɢDꪨ/"##|8޴+p-cV7bdJ;c{ cRny/F ˖~qJDA~q킬 o[x|NMNRO*u=a-@CN|@xC.X``"%#28}BK197iK2)P_y KM^>-<%.^-ڼTpҖr)A,%fERbȅ ܹ4~Ԣml\MgYV5}ypnz-,x{LlʂҘ kb.]wģܙ8՘@+R4otw<' 'h.$7n׀EPwz[ 6ҝ\OҠ-->1ЈpNkya.Ѿ0 N?n:oDl hn^ݹИhtgA3m%ON3X 9NjY_8:LOߏC8r-x! 6:Z8oVEqkt؜[t0(=_xTG)K{m @%SVs8=y%Ztۮ,ۆ=pE]0kBv9cn}]t`[TDSAixw(# Q6ZʃXm׃Q}` ueyYxC4&jȂ(VÝ"F-Hn'~E_ַW?n' b~ *Q8OL gƣ{d2k/8tᏹt2gL/;ARN[^og( &Q&hS'; F$ȸGX]\G&k3Um0ՀQ1A`=0l<\k8-|1T#{/J qsdjށ"gH蒝۬k;7K+f,*J1V2;@Yϛ`K SA! O~kS%"O&\#:+Ԃ5):醾Gb6 gy$BJUT#2wG:kEQTEAc }B"_.91e'@Gx079*8*%UߌBJ@=46E j[FzC_i*ӟ\r:hXEvﻔˮT#$Ϗ#da}`:|)-U#Bo _F;^㑰$Py8ܒC +>ȸ0~~{T-?• 8*$XdbɾVGj '-KБqgr`iۯҞq]NwθtN ;LvݟÉx&X=ݚXLsꬋ# ?J\d/J/5C)\և%U'_xz߅_dUŭB0'sBn42'('Qj,EldF)Q"R>(Z9J$Gr!9 lUđ:k 2Ũ0&0g,5T # e"W,fЕ8a aƎMyhBuP"^sW@XZY%%Ws 95R%f@e"H B[ .eZZS^n'kyQ6A[De?D$,*"THӁKJj=|gznfEDȋR6oQ:ùԕ1A*v ,쬈tTk+!:k?Uܫ`Og=IqJ>dpMߑ Ӷ筼hcHk,E`Np@'`@7M:KX9>]BwVS#24>,Oi޹znLb\PG':Nyp$PJb4Evr]{~sM//KoϹɲL!獄k`#іj"3!\RMgy { { [k$jUaPPQi& BJU=SNc0!IeSFVEaj8IΔW@ \6)jpe L1mIO2iℕ< dDҠecޓTS9Wf$# sG"4V#Bh8=,6`3JCS-T>Q d2ntSoOE~:BљWR ;,Qg=97(d]~X[S im,U CĖ0ɯDY.o9d@)aZmQ3)Ӣ"Q1*$|Crpaڮ<{(m9`h|FZIyVvҕځX;r.>@[w'*Huj9'{,[?VV^- .dnN^U{5+2aHIՐ֡Lrk#4Y9 _ЅĴ5O斀V={nc?Ո/Na6OmF b[Xhb~:nCǯ[ 7+3SH CV7}sP:i.LR#}t_)rWFn;Wo<X؉ ȲTԞ=` AU α4=W(: (5r +Ը£w4C4¸gfPWeEr<jZ71$.y\w,BU  ck-"5",R*B:!jE7vCWS˵8(oGJp{n3]iZ±S, Ϫ!VqA{H0R<0G1p#vegQ#+O&U`T4owvxM`zNkm%pⓀ+Wkڄ:[gU!,i=$a[ȥ{ǎг",usc):SmXg { 7{̣ʭCr}L b;yY~3_f<ߎf=v ej9vKm:+Z-uO-kc{g oBf[t7, q.3ZrZTZ4χL#er^f $cFVRM-q2Snn>ެ}"YvoVϮF؋Yeg_>m<?"F%滑gK)W_`깕`8I/,pE5R8%"֌OE08kF^VRT,$h CGzIoo|tϝ䥥мƺDUZ Χ>հ\EJbn*hqq>ݞ=W1l1ءӠ/1zJV<2iRm+x%RQU8\K0ymP*#5 Fi dc 4Ιj\SscuFi\IXo=^3#&pt-\ZJ#E%`aD4,_OZ`AXGrB$2 ì ҟJFZJTk #mnln>a[6Jg˩JK\V8qrAcQސ7߭/*]VIOw " D;W\9O^+q&2BX3Bgf6,3D?>ږ3H^7}KTK6{}5L7 .C`yG2SI&z ;"JK{L̕&t|1>b~ez !>!iKs >"/$"GfQ))&KcRH$2@ 9%UMRS+n ap8&J8 ek`u0⭑T+骀~eHnP6i^Uq6 P.۫` Nd'%|Hv떲7)T*w zzى"h;wȨ A KY,w!vXrt]\ԧ/U0wP %Z*A^K}*-:@xjaa b6W=:uɷ?)|+?gIMoX!ߜoFiRA;U\";~%ݳ5~友 0 `'BK.Ə$J E9۰U[5}7$c\ƣžHQd%i%3EOb-ѲUl< Eϭ&dyiEɃ)ܖz'W*ubx6f__ELC p1^&ZٚjQar|Jm}>.qw|+ag_g_LI=M h{U\)7fe9gm=@%+/DRb1ypJ%>^q "-4PD"n犎7m<]f0܂I~lӌ:/|O}9|N~WDڍ.r-o]3[k8SpS 9_I`pRMe#XN1Q'mdI#HgUDAxD(y [Y*-\TB7yw9=>?'8%l'DRM̶bO M[>G?*>t+m5N4Ⱦ4 zOX<,WcZc2?j5ŗY@r`Jh61S#Vτ6Ee;dMw $!cx<>`^3iF*D =\,4"O,С.^Y# 0vzUxƳ xT(#zwJfkһU2TWԆ"6hTg7 3VX!ZƤZBd$* )=U.F\2g&d-$;O@\lr' 2[xy5_̾rhjMt!j|̣FY툒k)`KFK%XQWD.o.AiūKoIŷ@h<3\mo %\ <3u*K'hN=p[Ң&D3ĥC^os=p)&73)~,ZS wܩSMƨV_rY0b AlonuF7mF_-S(?-N *z5nY98@ clپW+a&M߈ۿo75MƨfwLf`0p5ysC6Z )'gBZ ifҮF)!~J_q%O_h7V%WON_ͱ}VG9)e `ω<2 qKxn}9e*z<2C!_CD!wYGZDBĝD%$`gl% DQ2ː 9*vXYe_ uZ:\?>)4Z5w[fւU;C@B$! >Z0I.t58߼+IOM@ PpDydSDHX,1x⦫FsUaܐM{+H2rnq_EΥՠF3⸥`\hRB Sp%P,I]Zr.۵ԵhEq 4gUyn-ZiqǠ47 &Qbp>/>Tw$F6ZV;YtQ ʽa*:e1:Xӷ*vvcr@Jj8>oOQdnhH,nΘx x jZH5)KjHD'OHSThctSTk/D澠TI+ϟj#\\υu% Y8"tQ# ᫹w'4Jz$E (R ֲi!YdpSUEi6Ys2HzTUby17 ̦wWۋɿ^]p/f~:,cw&U%!emw$A|9xH:'g7y-q'!.V:ܤP*(j/Hw#e)sHsGjj#OmvTeM$D7)YaQY>xr#JHS/6DV픈ػjx3gw6ThT|LXBh8WjΩ-R<MކW\3e@Wau,8+{=Yw\矖gmo͛6鬙:ōU*IЀ)m11`V#~xG.[Ou1_gNj=LB`l=!6A(sIi0gq54ɣ6)qohK $s^u?eZ;TEVhn}ʐ4SOTmYAzc"$O2_7MeiF5cErċd([Gs,C?=^t K2IƢvP"F<M@3TFk^D!V@Lq_rq|кc&kOup\i-w?]==:wOm(}LsCαo>p;zF@%2Yuww9"ېPGtv/OqV$(e|1}"6hEE|1i(j݅t{})-,3 tԪK@:*lDޜqF&tߋ:aqMv1㯏/vЀ,|3-wO^8M5n(麳2Ji-ϸm2FכP˙V=鬬ᇝ}d$}ؕxTv9$_kENoa#5=aU{[.9_ke[*'rAF厷f2W\UhCoC5݄ Ab+bRb(8J0@lAB11nYMl} ֬AH֬} ޡ[Ju6;+߄)[inwOZ9Jy[NZJ|дwުֲA};& N:՜{u+ZBAwE| )*KByֱ⒏taih/cd^4Η*7õpx'`)~~vI7>cIyya' $6'&v؊p42'XQ퀚hYXݲ[= GNx(pFƌxt\r DSn8,\Chj pQTb|\WEw|}YZUkWS~SKbқ_PZiۃf67׋!_WC`m+i{3li|~= aDpWs6LMbM`xlD&f0Vئ -Lh8>H@$>܎>抻` 8udE嬴7 jŧL]@DiP,۶",j&4kp `$FD<98NKai\ Z=WLeE @s/n|uv~vш`9 s%8[m/C4օ*wQ@Ft}y LfRԴ`D-~"M{kljɕp %uMez@nXu%-$GNY7ί]^HM*Rr)~makJdmW3Pz@ID I :ba##RCL⩗&~f+0&VR&Z;U-L{Ubcˉ-uAQ6H<yiu]B+ > ,p+ DPcyӹ`"9*aw^gWgԕxb ;"x,U8eaVp[5 -TBe *B]l yTLccexL) gUEΪDRdQTnJk1 ?i֥]X %RbAscp҈P̂^%)`8iL{h+ XêV(>"V5FXE)`SՀURA jt "aJ:&J- U)e#`AybN'*FX8sK2 y'L~F5&yZ.R߄A |AaXzaQԿa|Ǝ6 >-RW@~A: Wơ aI4yzYZn%BJt`4qD'i+w<Ⱦw]_K9kw}l[NeuQu:6@ZkI$ӯQ򹞵 6@uߴ$ll胦8АR͡} y'qCNL/M&ڪT[a2m#Ho^'fCkIx; RGvHJI-@b<{ -boidBXţm3Juz(tOc66M!oE5L ܪ񺞵Qx3dHS/~l8 3ze(g0a} W 8|E0 BۛM3xr %j.KsI+]75j21-KBwv|>YM:G~ *WaCJ,n g,O:ᄺٰ|VXÞ믢ٗΈ] @;|-։*ϷO`FJnog|ar ˅| ?1_˭W8!D`0$!$E܆# fіgj'KWyTq~l =$Ej0Z ?V|4*m>5Oŭ-OG="TR?C2ˮM1#)u">*-&*t瓴'cV%H Y%(c}Q@m"CG^>[u}%uڍ<Pكװuv} 6 &R94A,+Zzk3p1w2.wy;rS,OHY2$=4)K mc,{vèGo>Ⱥٷ3LJuQR*[. w.mӨv&ESTGpPH; AM4D_Y`*B2"5^W q]j>밹dI'uظ^p*?& 8QsÁ2r3uƂ3$Tn?5d SKuVlyѫϭǻf"{\29=6[HȩQ2f)>c8R&IdlILK״Νmti+GڼQUӭTe֜4CurU!b>HkxMSk< Dь9Q4MdxOU*o۵2E4|Gsq55gW˵Ys rcQOvKuC=կOeʨ`Yo(s~W^9AD!qoʿ%Ir]-߄yڅ(}-e~ 븈Ѐ{t.oz}` ;k([?)ήâ7著w`lBBZ׈ӑ(.2v!"$iOf9x3m`]`:Zu;t׋kL3O{ģD8bзf"85BqƽuOa V~lH d;C\nU=>HiDQC:fQb^ M;>L"~hy*=Z|>YDh㱏\Qi6T"-Zp\p lyƱ΁JofΥqEb|\ƺe*]1j&--k ~/fש8٭[k$ 7IG c0 !v Ftꌙ'#;/w%zїXJiY0*"_J[Ǣ4w`Z۫[FPD+!>YmYeaH59j݅P1,2W\S\FS&2Ods6TYy.~e34F\F B3(0ᐨɡ@$*sU6 Qo ukn{6ͮQC D#=7sfeʶ$lꜬ\ܗU#v#0>k:K[lYW﹘"ݔ CUIw3wp ES _Ώ|.-9DZsXG߁.^5>uZ7tyŴ ƕ' 58̗G롡Ys .{XM 㐏YD5Nygkƍ yhZkҘST ՈkL >DqĚܶ-sim )Wo;ۓ_ ŐP[db $dsB!rDPcxס\˧`5Ȧ+9;:IcY0[B U*e#X'>Lt5uF(,mdET4.]}ljsX\Vr\bhsJ>tC6搥pwtX?),lԻt*JH~M&08QA JZM!qG/uQ|OqU&lP( $ǰF+?;ZcNfHxL>H EHQݣ@vJ }rB8 0U(/T|P42UǨ t2'^R@1K? ݰb sO)~6ZJ&]ZD~ZW񲚘+P ɄTg_Q" }n`ÄAZTDG}I бgg.{K%>.c^4y5Sy9B}kBC",p.ӮHQ:K[n!sL4&&_9Sif(+{hA4. qގB7aْ!D6 B6Cw$KYcI !ŤLRIFYIg~A #$p&:}?aG#xQRKt|!\ :Oou"xZ؉D}ǬavqJ‹˫o_L mdƽѣz}_­ϯ_|jGC$)LؤchoNSzS8+E;D \qr2vEbJF=k$Oq770/7췗M&L͠酯Ln{o k1щǮEwێL/iWfiv?Ao04?n/~z_Zx04ש JY3ixe58~S:2$9׻OZO?}N |-}k/;R捈ⓉOALtL1<1@ zvZ`<:-=6eO{ʈ7L9ߢ?@M&tTn:ˬ.|=3Cw{Qֿ߾>Aݽ?~nŋ0Wq+6p:l-p# `)Q6L>'el)[4~A Rz_~8)\B;y|3׾_L?_A5N>HEݧTFx^]dndwL0?U?wOYIIt  ngښs@L:b`vtµ[U0{632H#y#|E*lYXn7߅j,%Msz<#RMs>#f^O1QHЦy 3`,WՃ|; ֬9U&\4aP5\8mF ZІR&Zc_C1Lڹavߩтh$Ĩ)Y֠[it)? ؚso_msA(/)]oӼG^ce?ǩ>zۋ IhLy Sɴ' 1aD#i}t|vh08>z.G~qd\G2TP`8 TaFi~da#z>B >B A6 9o+]T\@Siu>Zyx/or I{t]ѻw*ur,+1 4o?( SE)bAୢn*ڛ q1hr5&AM)[U(𐧠DX~DNs7{qsYo2ݞPIMk!v/, `yG4QhFwꁏZ,Hëб`vN9&QNf7ӥG+\ڣxwm"<'<'<'м@[0/.k9"XX|m>IZFftUH+ Abd¡L|e}jb" \՚JUaa{\YT}%~}şok)$ }K B!J Y*V'"*P D?~X| ߛr|&Y8Όa4'l]g:$ d7:_n'KWq<{u*۫{ f pRg^ZQIҽwPp*vEnb9<XT `T0 2 Qz9Lr/qkd}@+Y??^wQ|빔[6.@eVCxO$tn^~ ͇\L!q4+[9P4%( !,&)ywn+%;H#>[E0A|Gr`) ;6)3ʬMr; 9݌KS\:ңر0-XrLqUp-$)_P`vyKW(|B+|I ZsWhgi* ӛ|aUe8,lɗTanmw#4~@5hV]LWl;|lҸURrV/IMZ,=N^skREabŬ #.$B(xB[F("R{8ig@/ǣVׅԘF$Gi8yv/K zj:WzѥcJ m?r&ڐLiLZqf,9q:/_^=)YBpL GI3`M D7.֭ Q+h^eE@!B==~I= ^}h\@[ ܺy,$%WI֏bwڃ(=|Ǵ r&a`+Ş"{ sS&8c\ήgF(Hd%k-[eL~CA`_vgA0Xbs-T*!O욐]3̿/YC ydO?bN$eL*TJTYGM'n 2YWgT܅CB[%B"߻j9;Эnթͻ6MZ׆bN#T"WZ0Hb*ۘû8Lɨwud;ԍNZx9m5+_,84drjnHZhIQ/ $E6ȌY%BIgsU]K5WJc87iRka2r8[2sF*x%,e-| ^DV4þ km&%oOEJRK k0{H$&»jeӀm4 R(B1`]= VbguaZG~@lY`HdӜ=ۓ{*D|>C}Vt>A {/Np.dmHSj|#ٻ1[vZK#TY[{8z#h䠡6J+M2d<"]GCdC&]xQ[ў8/GLjOL_ŖcTAQ) VCuhX |[b7BVju\(a< 9Y¢t(+q֎-ݱCXݰPbdǢc6泆G=I:wͩbdS{Z_FG0BeaWm\ cHe<&wk]Iׯlz8Q61Kt{Q2͡1Ko2Dž5}6o-t7q:s,JT*K )rH ؙ0Dϖң u.4,e ]O^c62^ϊF0>O(ג؋kێFqud_@׈S'YXDAlRITޱdJ։$1AZYkTo10bF`Qԭ tP 8^@X!J"NI&kȯrb{)&#;![SKa`IlmчK;KN[M<ޝ::ub1^(2A'&]JY) (c`*?14;+f)XIaCN=8uR +5Y {Xߑ%jLMaGD cBh]ʬg1U}sUȒ)OJ+S9{5?U8}HkSkm-R<Q0ߑZW?~&w2PM b~WH}RxsYn/nO?,rG!Y6r\zm='-լ󾁽A 8[zUj@0^@u rXg*2؅iVrݮ5@#q^ڕ3cBIs^?\A&l2QqԚՐHIg=钼5sb$+AU6@xkRBI#!+2ҾVN|Z!`eH[zoob $nc7d8j'oعbkG9c8&F+R :fP0۵eXch 4n3tjAR'AxsQd5li>Z*>lckK{x_kԎ)V9qfU#@:}w`c4#` JW'՚JwޮEʴ#vW:`#}mg2[?7$An nBޙl6ʙ(h]|b˦gsEWⵂ씱)Rūg(MR&O"f}O6:|M9Z}i:p7_% 8L #D=_6I_meGlrj-x_vB?;#7|v&ZFlUOuS3fg}}s9锝!Ա!ydYd  2AI[PS[͂鵋ӷ3sN/lx ڶ}?w6fLMG>Mc={3=% hHɠwDlksI~qOmxN}y' &Nμ RIhasCglUJJDz/xyc_^vӲt5ox6/CGj EH?x:D&J:S F 2$ BbvR;ZBIe$sP*Z.}8%_.iQrNW3|ut6_$j^wvUVo^(6⷗]͎'篥T- y)G潸bX p{~ӹ\束4g./O>i[B)ZӠtsD *ȑ?y1٧sP9# Mܽ0 ,}ܙ݈!Nxb2b2[g  N5 )BwO8׫[tgGpgաfY %Ü0`㕋VZ=f-آKN̫n2f?[ ;^H4u1@ԫ͊6^:;ؙjSAqW'GkVqJq# Fq4-Zָ76X2& 9R 6|)%q>gT00io=wQeEDF>΍U^N.ԹO|-a: jNhR.阤W>1ZO'  /dk0)"c[63I-ec>|xw}?ڴkIEhg lxF6ƶ_}s:bBq7 ;M5Ebs Odj7hzm^}e^Moo?^^k3xqJ jk?* @3-~+-:fH?û/SRXU/DK$]-\ Ȳm'IMVs.|krB`#+ehExV|ѧӫڅꄥ|{aXZEώ®ӫyEeV*|/y5w]i?6!_]¼I?pĢ'K>"O[Xͫ~1jk~̇\PNgYNgGm8DB⋓Z|qR/NjIn67絼-8{ y85'g_\9KL*a#S@7Toh'X=MVӥYx1Ϯz]gi@O>)F},Ե7/AH}Q?̉M<9h^֢#'`40vٿ6x-)H0Ozp,"Y Zן۱(SGUuR ID{Hϧ ˮZJo!y7BVgGwO创$ ZKL>_JUk"Ca#9Mi"a`Q^lXHc0Bhg_^NjZw] W7A7_ z"@"SHf/k*HȆ$۸Dzr73Ff(IT%$EpEXGV_Bi]#Xŷg~sr7g1ʟ:Ѳ϶;^B!{6AJZ t93RIkOUR .*si1akI*Bdt0%g5E) JR'.)( SԆUl_o6>p=uDC/eуn2ׁ_Ŕ>lw~c*٪u&o~[ +/'VXkOFZAFs1Fd>;M:ŒϙVʋDުgy۵3U$#H㟲ì;'1^; a X</r\Hf3Zpߗt|lwO=>l ꫣ4͟7G]gkBtܞtz~_]YsF+ ƒPD?xݚYǺ*m6!{I@X #:$M6g<}q}T!􃧀0=E}LH >§$&jE˙iWn{ᕀmn2'}OZ+ 'Z0Wq#=)`# !RZ Ov ~ǽHbƟ:sB6RS6`RRa?-i Pp!80QK뙠)eJDd q2m3cybF9oK|&ITRuz&o}[r[`X{\JcP2}9FEgScv3GCXk|55=fڀt&9:d;{r#)I HD4v%{o(I>l(Y#=Ev v(qLS ܯTCqnQsP0 (J`:@Wq#0 c0<{֋jA|Y/::g2xz DpUe՘ϊwZJo Td\髤p'>~B&,; Z*-#"h"}`R2ʹ%,g(';v2H-ho)FQ]{{/l@QFQ}39ժG^a\8n; ~+9 U)'523 piPPbٟjR`S>f hQ;64j%Q+ _Vg s_c6}E6rٮ.*bʻbbR339%@L @L,I1ʬ *Q3q3uS*A7cE_[1ټt6O~2INc bYLL8lD iP7ՙ; ?ys] ⴂ8.N++N+(z0ܣG<ř kqG0-NJau&,P= lں>AI>+)]E0>qm_s޲X1@ґ(2}ǁMv1%)i i+ANP 6@+!?s} =OLQϼ\f&Z: Qc*;ԭZI+/9Jsɡ]_u'6lFqdS}E5h嶷ؾ^"`:U^*K//TBBFB:1'j bkm:3藲5F꯷w?mO}[pk)(DžL־?  ۖȌ õ3ifPiF5A( ,|ӌmō>zŹr/x"bJN8KL\se9&py[g)/to4CONM_3-Z&- `[Sڥ|wjv"KH4+.Z`:7XRDsH'>@Flu8YiIb=z/I)rE:Oǣ4҅p85hQ_b2K#gX V A^ e.#*VDCkSELp ,NGӰK1􁋡 EȪ ko7PꞞ ' Z1ſf'CP>.^o~p8gI(5}_NGX~ìg3ǡ2~ȊTenE^PJxŠ}OE"ɋ[Bn4*&m@;x#k5(E KhMύyƶrbQD` -fOZ211ɐ13hߎ Ӑ9v cSډ+:]9uKMLN0:h[3xʐYYP} KҪ>&^eֱ/p{;bvh[rXj-!Vs"II Iζ`tqt<8 i9˴CXHO3pC-͔ I,ч~qymëwA#p{ *,8qل2ظOsT?cPǒݞpc/,k-ưm0oUŮIaK:Ytƌ7 Ó.1$UPJz)Qs<)F:iA{g!)Uuuʭx&6BJ%î։l)"sŸˎ8Xw"")g:U΄;8ahvz(yjr06b]iZ W)lVQCR ۇv*%(Y㻬̇]lt!?#whhA맼^L:ʼnʈETXmXI&exɄv>f eh($EfwdCtPAR!UlX;ׄi29(.)a%^W &X.8?dM~xUviJcYEBoP[2$x[USbWpBmYqL)Ԃ/,rc)U4=Nj_ \,o#㗩Ilw.܆l4o||]݉Ȝ@1ZFwuSX֥N.l𝁼e>1ǵxICy"=oZVO8em=Ǯe7=YPg@m(W/ak@h]fdzJ 8J.&3f&EĚ¯GFu?N#d {~ظBtޕKDR!D&:pt3p]A]ghXVz {zd==V:Ɩ[]u&Q*Qz‡>ɧ0{8+Lfl|>'Ӿ@lUɚ-" T2op8<'6;~qx HyLMzݝ6PK ;=Ki;{Bkq8'#pj;ĂTp."TSvCFЁ_[_)Tk.x]cM SoCθc#A@1>mKm #1}bJg;;4?h ,LIae>!5*q)X7hvZ–mfR,^;ɤ# @E] rM 9uVؠKQAc6lB{5 ̩F+ļz~ke1rEG8OVR|k r/W @Z- r0 @A@CY W΁E-f"kj^fKYkο97ObٰTX97ag9T'g{aj;r\{s}};Hk8b_]n8WWD@,Esk`80{pB*~^u4 tNӉțVԵltOQng3,e=Ji! ):Gj@#/>=ﳉ-5H<`f"HfgֳJj&+; :]5m.Ej|<5 Ÿ>=jE?>&у4A}ZI$2/TtE+>H8f:sYW$URk)I$)]7pMxGHw=D2z-~`m^۔eːI`k{-4n&r\Ss>&tJ)i'XHdNu+ RqOçPO =1P4a8ϴxغʿpUloٌX ϵnFLX͢]a ʴÄ1k2JU2p)sۓ\'#+ej8,, KtC.Ш1n)JX Od#\w( EdbrK|>|uC3So^l$d੒o/ pgo&q1$z2Զs iMZ3O`+I:ykxEa|zUO2BxLmzh{3*VC lX3͂H1,#䬗[ .i[M6" ˛Mc~+qAa/~o8>e/Vo8{,ikӛ|wꏹh} %dYo+=7݁B:4;čw?*yQJ^|gr|9jixIk`BwyN}_`ٳVD0;[ҴHRD% FBvx reɆqPu6 *74땭9!fQdθc.%odtJ>#2V3#o^:V ])A~#gf^ mAiGd4.A:.EFLZNx+CD)jR%7g +7ǵ5֣MA!Edf%2B-!|<8)]{J|;-CKE-qt7*Z tA*EL>dQ)C3 iz 9ttqj~Yp22Ec>]^:iy{HH[L"T]#daI)z6+NI7J=,e FKA? +|{sM.溞RTV+5oVkVluݿKU+ -cQ :RVyKrW䡪֊v=CZ$znJwyΣt½{r}lj5j6zms&qMp6*>3v(q_^7[qkru͇8~\Ià Amt; ^qwxO:Bl(26c>sD^njr_vy0;/;ؾiV*bWq[{lmpuJyzvanX)eK3JwX-ֳT1DAђhHdC셗 D4e2Y_.hgRkSZzZs\:#7[NdZ&:B><{BG/69tusup|!*bвQܦKe7"Kcqns;|7k5/).YM)pk!&@c.#뾩dHVyaZʾ'Qu_}?Q?-xn]zu s+׋N:?^}c"ηPs,@]Qx b[+u͇ \1Fζ:v١Cŭf;!yF DUasfu";MV)w챖sf;맛/mI3Z59x{hnhxih'@W@fqKcɖ%|v/iEU9Ҷ.t<{"aU=! f]Tݶ]1eJwg_܍HTr>Ϝ%zzv;k(:Ae?M苴#9mLBatgEmu9i)vBȣ6n) | 'n0 *&Q6AX`k_NBȳ42XK2q4K`2&!Wʼ~DΊ{漷> 吗d2'HJlHb%+ύᖏdcg) ܹ;~.穇.쟽Zw$;Y.H {UJ )(mwBzXT>p)IrΣu#n:Я}h1*Gp x;/~ (T{l6 2`.G[Ք\Ŧr/kI^"uJ`gpjiZ+W r\bE 9i4=n6 2˖]jjZ\W`"j2eTiNnxh׆RR$`D>ېNN/(Z ؘ1`g%I;xDc2TUnSNǢ-DyKF3H6jg)&RJhI= |ɫoU辯J$}PtH`L?!reI)p*G!"2"Ŏ&hOӠ#*OGQ&\\mc\VM00rRN !S ۓ,hK̆1%L'6";XUQų=`m9% JZgȜIU\:@:OJIt(ea }9ΪNt16 K4 d);6QW,((Ɍ9gʕS(M΃!s C81497u\}݋.ښ}{Q?hNLĐ-֔?mI /`}b=1&7#)SvX%j|x-vFp{ubhd.M9PO-1|wC /txUTHوʺ#+`gx2 x嵌6{ik_8 Kh(3Ϸ?x𰏁XeW{W#Nj>!Ф祇A}TQ;TVOB^2L*r 3!S-bH2!疷 JW EYF?m}ez~)JJ(OY)G,lrYXgD2(Ŋ^8F% sP x!g5Rz @ 8(\j:zE98-O}RJhxs(r(/<3M> M%Ͳ.?ONO˗3W'Otݙ9MFј6&FC䀬ULttMwE0d>p'g_8XT: o ԎAD-0XQkv4 "R㊨m> j .3ѫs0njj~PE[T#l3+1PP{6 Gw{[л-1Ab[B [CGok-^ބ}<==Ў(= 1db?htEkƴÙغ ; RL>)oOS[ue~>X.Y]"u?&j/G[K(!7\ڛ!]::!) OdR}&3K̵(7 KW?^{ 问]/A9M`!4HqHnL@/vΊ,4$m|Yx||zkJU5oJhÜS̓ S\rƝYOdeOV7h K.x^*l__{@DQɉDnƑ|[F+1`Iݖ-{OqsX49o*ڤ@X>T8p_ɲ=12JjyXaQf;+&߯j rU9hEM|}sqV}FZ鞮v/Dʼ/Ci3\cy`hiyC~E)yNwa;қvĂlݔy}gnr| jzֳOD-G1>AdP,բ 7!=֚ af7d,FpCǽMɳ_UMXxj Ҳ ar@j|֣O ~ty9b0%ӾHL;,v5і &]Ibp95a@!@0`B$1*2y*it#"dc8 WY-0oZ 7@,FzmʂFǜ,ѡكRR 8^t# ,Sbݢ Ջ$l[,0~Dk/|Ld0i{f"Ff[@nx*2FՏj5D4A6Nm#Ӝ!Ӽϓnz==je+Ӄpy'7w\vdQKvaVzq ]Y. ۫wEo(~{J2o.JE~9/iALׁĥ<4r*\("IƃJX\}w] nddTrd t"[d1=.E|o9r,~w);+!Wpi͔RHrɝ{ix·_Q/-%-nq䕸 Jvz *FirϗWߗ5i*6J.ͧxâ0Uw?xx¯>uX5gk 9Q,VK} |aP' L g"tauPkO/p O@EϴNd2!0K rr@H}vIj9Ի޽ֆ\UzCXq\p+y*ݱkh.ZI96_oοohj7!+;.4vʛ>Yu1|chh.(",Eg9xM揢P%$=E粉NpFZ KzIO1.iqwAN v4 6 ځ xہV]ۿ ,UomPWj,:r RC1zGLpyKpseVqZSڊ*ɾn%Ҏ@i_~`-j'QWX6nͷ]m q|ӏ_8{Ofs@uP'[,*?[t@ug-%ȩWNF]^eY+lȦ.!ia >zmB،^&uw *nB"u̐-=֌֝MK~W | p J(, 41("P!`㌨Dk#k|pN6Ǝ1}SI+Ixs`E#)A"lb2IJ#,4M f/UQ|%oc>v= ĢTLPachLPk p?9V|@)Xli_Z! *|/&4PF[>}6o_= ɂP {1JB!8ND#Y0'u(0#sIIAzT|At67y[5 [oCL8Db~׳{c~zk]=LB>=s4vT M(F8 =>p Gdb1XPYR0tfNx;(;jd\Zinz &7STr/UN'qe jI|5=6UpDW~p%" ؼ|b.hۍ?_S Fx a$D `c4 aYVN3\` Le"P\zPuBMeer І!rL@E_*NkTGHbso>e6T f)ڃFw\I,Gr RRcJHT`3Fd>f)L*v'>{k$27:!}!1{̷+؆>ehmn?^x=f7K Orkfa,8<# FyQE,u(c)EO>ng]=yX3W.z97"2Z>}7d#f⿷[uY]ήLgW+~:-P~٨rݦ鍟54 Ð{muM2+,h(tw&r ` w|IƱup(wQRt{GPLЁ0F[ Rւ8ާ*Y,} { }-PLRÏ䋤1c0CiqC,dk?:ђPB !ȩɱ>l۴Xt-vm6U_܀~;5?]-.PsǗdx:^\;!gDXB$g74]rsM[S rSu抉4+DGdz;g,Y΢rh3iAL8}6gp|tRs&Wp3tp.ŇP̃_N/?<'䳋/^pq i0N/˗/ϯ|翛Mb=|݋/^|._ru_l)셙2~7~~9V߽QvA-g>~`Li`}l ;5 Gi|2{b wożtr_g٠^|DaxMΗ-p:ȁ-ؓ9kWC?dEhguƫdsG< nAS*Ym|s6rD8A0@vCYӳl7W_1g.3<{Z=z9L!ӛ8h"8'>f3D.zEjil`્~0 r+0?0=L8yp\z|twpr5I^M?N& &GߙG^.R V,^10&pMHmZ`0oYoooϧ7 2_ߗ?}/ζכLX݁X|!+l:2{?ov9y&q.ܾ@6E'WKY(+ (N^>[)PPAXμ]:._rĵUw+\Tyں/tyʑ0e^Tb5kۼܞx}:y]zvVY\#E:XIJ/ :5JW 50vLs 92]~o٭yU>[au< )7/NXlˬTϮWJLAe}Lwq P3x.^)`։+P$jQEmwhzAcHʟ8us7XsiҸ_^hs3 ,X #raV)YQG5^U/֊LLв)P;s*{FNH [l9O%;is|M&C+qNQj%C4ؚ9[KA HŜkoƦfD[*E6llhʪY [?fm0vAof}w|^} P{3WDg"juCe9 ?+ѓ26xs2VDʴZj=V+&*sh%u@עմ4PkLpMn5lw@fGAX HI "I$@ hLE0"#4 De&u2n*;J Ռ+%nYˀG6Uؾۖ0f6Bs j78Df<JsˁVCȬ" /HKD> "0$ȗ1C %QU^[?x"KEUR{V,•J=Lxmcq?P%"tJE/EvStj}}\LȢNQ}uũ䤜uU~ny%%tdϟ6/6d*9{YV_!g?2ŃZen6M~)#}Qx!Q)Q[?̌ 2ybiM|J<<+~&m>Mc}wE LFv:0 qhPX %8MRY׆EpGK c [ubUer1*ZCF1fkQ }O! `[ c0 jXXy`a";; S%AW`XKJ&k~=^ Lqҷ! 9[ :qrOVPBCVF"ܽd\U&7Q((;j(/GmN($V PŭJ=BaY_*T:*p G[X?'4֫-a6_2p 9.yE5KhR2IJ#,@&$/{ƑRnt5n$ i"I91_F|HCeɰ-T^]])2J m`Tzs5Qɝ.ukIX|Qi4=slh܂[<ʨM-nB#8!z)m3]8CvR`:2Kg鐄Gݰ[ea/ty5Km;+F Z Vk&j+S7fh׻bYCttd%4j^4Ҽā/ޝ\ڠP#Ё "es8O6_TlwDPd\YUƪ'?,/j)Dn~Ѫ pgqq κ/ѳFyeeiAUHL?V{z0O.b򀗾{5ҳۼS0ʹ.';:ӛtJ̈́PQeXNRo,"i`9Dxh&.}NvHcY~x> zȎ즗mʭ%A|M[bWm[öa}l9K6^5ԅ^ yft:䜯-y5tkFKjHQdnpreqo!"9#x5>گԔ9*f 3%c=%')I 1F-x-M9a4^-.`OR1qY{.*ユg;meX+w;gdU%˛iqMNSYmX:ܶW9@ 58y$Ztk^'4nA׃TV|G ժq%FsǴ&;?Z(%SHѷPMNTکq)Y@!,9gZ%$ĸ\ eQT{db-"K`7o>]N1+Kri$uzڃH660j'u c-G LW~e9~$z5ȤZwR,f!AP F"Wr~υ(VTq%2hqsp?J`u5Gk}0 Swgz#d-j>:O QF 5UR9hBt>@"S gࣈkBSrjq;]?V!lֽB4^3Zdu@ ㍣Vi!?HKl4M,>ټR; YS0dsvqNkGj Y-fu~}Ϙ(NnG$_5/me7[c>csTN8䙳h ^,J7[[ Ng#g;nTu!ϜE[8c"{nm1h:mtۧ;nTu!ϜE7Wz͒y4&˓X+[}u8.O~e'X*!o?$ZPI"kDPAV k-'-]\71WB4ԂaD$M5!FJ2ʼnEF)1/!мp['4llFVl|>k kM|'[Qkr^N.>*߹\#kʢlrɄcΘVlvN4mi\Ol n5V}$KX< &;hVF]Nh84Y%暲aB) kU= *xbg=ÅmIe) HgT4s}MAE !Z )UBA=}q3834)5l 68Aw"CYУd!ƹrym~y)b1#bgUNsZWh{%n6m0zPJ4|ZH3k]ui 7x~#WC\v:5y?ʳSqЬ@p^}8*Hʒ3QKVFc";uM 4p[ -Fq&)2Js\Qj''& $<9k\}ܑ'H]0Q82 :X 1[ꌠsgRRRRKS2Q{sˀ:`ҖRtϼ 5Rk@.%v OC\ $3s^YpD ȇ^ K![J.z?hV'_%,AF-Q 9Х{8R(e2(Jn*-#ȍ!dDfDV^B|>xbpd,2NQw kK^s$BRCeK}ltZ6rTR:X&}0 ͼ"m$ yBo RQ* gc{BdGh r =h@Qpt]|BAY:ӠpFpQ:TDhNi4i!`Ya"G8⛀vq;ڀnuB{hjgzuH~s G"l!񟽰^M/vw_ؽpH@Ki;0bAz.!!o&91G SER VOQ?=O*\-^ PւJ]`p0qn 7H} , QzP1L9!Np2")9jdjj[RЁ(\F堔[w9/h!&H4gOjZ=lPtSHՠMj87-΁xmdg}Fy aC2韭:|Ge-,5`rZ!55O2hI*ho.]A{.N)~/?xrr'.8of R&(Kl>Ծ_ R5n+9HY"nq!' xm~o>8o2@D:837L ٿnOo(ɿ0ϼFLG譁h\!2e9 %VK>! D4۔aeՀZpmC| lLbU!oY\$k­2&Z^EE b`4<'}DJM6"БSKY?DQ5az\D8`w0Kj]j{4SoCCk[.l_'i Y%9n | ֓ +i%BPLn}3;r|cSw70]܌{FZ n)л_ܮs: J6#Gkpy "2D% 6X|4鿷~7zpPGm@ ݺ>\3ΤBVߌyWLjՄdi=IYcۭԿOX+!jG%Xbs5bӛȻjQJ6=պy]qvk4=P0|t?e G`~?&5S|T^ ~0ؼgͧ紑g|//B)غ}g==,a^ʽ?=(I5}:zm  kpEOYX3*Ũ&P(^\29]= :+f}Yf.晒/:׃2+a'WUhؚsa 1b,܆|;/W;免U2ǵ)ؔ`dɎTQդƂ̲'2 PpU  vJI)"{ vJE@:k҄ 5-aoґaK ‚0B1H0#yz *%JcJ0N0aj:Ib%nḛBjTJFűf[nF&~$eY`lxpI@@Gvx2~fBC48 _}~ jWQq/Ty մDqz!w_Eo6aS I8?Rs l˙rS2+<%9ʹ72T05`Y8E ʂ ڦ ҹN)bn /B0+SvQeNu4 5ws͆pŻ]T-0")kp{{ (2oxѧ$&X1hv (xrBW~?[lO8ΖF/^:3u6!ٌDVR^4Z hCKnQ;8"x+}dzr}DǪ n˒ig_ n3u|ի2dsr?GiAފWWٖR~Lr KJҖ彛ht[xi*ж= ֡٣gErЬHPNJF`)"v^c\-N;W:\RrL{n+51rG55Xs=GjdvkpK0)j8_땇֣BM)X`X!)M(z}qQ%:UU@iFkc8Q4ށyd 4=;Z ]R1od*M;4[2M1eUG[ \bAB󧋋+Eͧ|ݼ~r7 nl1h<}Rs ƃkSL{2a18Ϝ%u_Ӊq֣b{gR"1hԻ\Z8 +آ6ٲfIcSȡCY<\F[V4+G6GaKHu*T դr%; ZUd_3%[`ŚWpI+`ײӑB Y= lkBP;9ԄoQT[=Ls$aEp!!"Bf9ӾJiq3IS@ ;V+ЦNvglOW9u /?LN\*qRvgo/~zݹz !1G{[>x%IVDAe c{KVDYM{'E]JDt6a3@wr.kӪ$cƤ"AgS `}}}on{MAIr?o&vz@zP@tbHWdbiW :B[ϰpϙ7FI#i>(g51 k1\zJ)Mz 1dzB2IVLԂl"<'i{\qz@}>+{:gC $);OC}jMiEG^'f'pbI\TpRc V63 ~ 8ҏ<6ܞK/yrza/R<߇G(8)wfфM<&oDbcC\ c[IbLI ֈecXc3 !Ke x: JґuLAFFk 8'8lqNJ8E\k)d@GJxG(:QxayVT")&A+N-ZLDT{V R)^ N0%:PWY/4GUR8 7/)~C$Rne\͸cVw[Ekn]ɒD$1ȹ .MyT0PE6 s9x}zs+c|6'bEӫgSʩR]4c+(D'Fu31i\.rB5?F<9~Q7{T+OIFojӭh4)nك׃87^n=nwS޵Fn#;G,0" vs60(_q۞,Rv?L(V0[-~zuo]!Δ&Nc".Bu!ീRb/VU`9˫>8=Nuitq|;;N뎹Ӻc1w"{NӺ:mX5M!{'͝x{t=7 v?ӫv`7S`  . t^)hݎ\Ѕjb-6Y,txy:JUH ͣu}lD%iigBQ{(JѡuK;5f㌑wB4e0osqy33p` " ⦩vZeaY羷<..;I!{ڑ'Sx%$'u ; "1r?"뻬a{gRsO'4Ue7WY)bC*}!D7~2lL"ǂ}X@{0Qq,<͑⭬jQG%HvDzݩh)tFLY@BJYqW1DZ+fzdE !pt4@ CTX "V 2DЬnQ?rdh]7tPR?Zj _ӹ2Z>gO[hg`; Ҵ΁"w@և8Z ?'y{U>wjCW_bw))X{\}(Dmf%' ֘0)W(-LoǼdqm6/42y^0{Y٢`e]arAZkl'qӟC SǪX^I$mx[c(,$X! yu3mWIV:N/t}wg]WGW_x>teᕻÝp6+?<']^I4JQZ8΁.JՁ{S9_r%CIFe;Ty^Ҿ]|./UlE}o5m'KADinmɭFa!xCr:"_޷l[95PT`+ dQ9 6bʃ*Ğ8$s?S"kTj0O`.Ve^IXBF%%1g3U mADDy;("1>Y Q/sL׀Blo:_lxwCp, "ɓ߼/C`t&;?2@{ 8K#f@4MϐY׍"m"OۜoA'7@`}6kF!검$Z{ !$mjo 7JJyzz5 !쥤865Q)ږ[969~N[r7ݙZl1@+u=dJvl,7Bw{,j<{܈%ۼ!APjDSޟ.0\bʲ9YDe %boH{tBOkcrW:=똫Z$biһ;YA3YJ:K Ђâ*3eT#j{VM9Kf91;7"`La†g ;TIWh-H~?R ^!2"VÔ<H3(cVs7? [b_f`IH6w\?PXގnÛv3r N=+UT`J3e_BR4ؐrl)Cfk^_/.{}^C)4PZc#m9=~mQ97<`Qv<[o2٦Sw!05OkLz׳Oc5a'{eM`G#w2r+|ȳ~C";Đ*y]HfY)tD[^>Y@zS^o 8%W<\efxy 8UοP,gR keǛh4Q`>e87^@:9@gI6*{&nvH SM)ly:2&6y*\_~< wqwk??&?/\v&ߛ-z[3^D ;Sϖ@yN\Ÿ-m9W)Ec;2\ZāBi"'P=zWrzl4E#9^}M[)ݒB5Lsʲ<)/$kZ{H0%KtE[1-}Ÿܟ}Vm>𞂯v|`x/;7ONkAS7^X!SVa)0TFT<e.7Vڡ`1*^}w* 8y" Sݙ^4WEv,WPYTX0PE|H&P2XLS C VPr-}%J.+QpjQbf^2peU .JhЃ> 0]PV"};NDT:j'hd2T-:a\&CY&/D$)NI;ȉD!8hڰt*-q* X68Q8X9,FRM !:6 ,=)^bkdi6Fp) J`tmUib@0ڈeiot }E200[N[jA3Bs !@/VTMŤ* %(Lth*jwz:H΄Qe[2ؿ3k˜Nɞ&n#ȁ09JIA)R&/E? LVUZ:A#CiZcUw^F,")"X h 䵲'`U=z-J+8Ɗt` bJzӅr{uZ+ 5A3 G}v{؀V`ARZ?q|q[G=ӱ`BL͹bQ]o8YO)/Ιb6v{qABJˉO>:'n4Zh9ԛ>zSrFWbAb]ŠBi4:Uл; YDQdutWYNGgp7"!=GѨ2CjMl0KxkGc*n{zSc Dcž'7cE+u(K--U[#Xg-MJ@:mG "Fe ala. ӊ5fZmOw^[(3wv:[ZO_<(ICwGT^T/,{U]]U]Ž N`""\J%QNpTߧ+ayC(h ?{kLcx揓`k{8+闋bEofE!\܁ A"闛ǹY\'WbL'~>8wh1%K_AmЅ$V~2/ !NOZ= dP*-LAbD}`EY_j1~.@K/RfjS6p4nCyc1ep>Dguzp!)1DMQ&= !Y:%q(WM@yL5(gDcY*c)r8.)ÈH 8xJO3h6_0OwB~Xv4JZ|`cJxӭK6OD_)iRFՠQy´1Jᘼl 6zYi9npgs^:@t,zLE;W"Yb]6kйB`p <`Wϻ OO`^1Cy0> &nnWpj?'x ñ(Әyђ5 5 u[7V7*Ne?O|6O1gpS>5ꗖH~BH\{D"QBbë]OxioPtTc@estJpTN;gP;!8*5)/͢j +GRkS| ې[NI)|m(jk=㈢j6 <]Q@GHaسd] B{`&uZnwrP^RAX0tL-S!RU\( e#RJ{. 62N1H6o~HYHE: HLpE ՠ (_}a7rEa L SF0mvAmQS^ﻂjkݙ{s,?mo@ē~.}+7hRCϷ>BmOЧmn~0uCD aS`G YD$Kɦq8FR /'OXPAfJIJA"zھc& 1JO*P)L7Cg*X{ϯ+q{,g?i~!ᨶw&q/ <:39LǧO% 0%xű1kx 錷wf~c`N)IĚEXBmץ~Fi/jKYT ` S9d杲a*v*%ǨqMHR՗t2(g2`s+5~l= }_LS>*OTTnDSUMԑ{Nwa٣>h\6 שyZ>,I9.^Mo-9d;$=A3t8wU~4@{N #H@J)FNks_)'u^Uf ИkmfY1;R™L/uPL5bA"Q'1Q1PA;@5OoY‰c+Kx%(J~R]v2M9*OQەe:/}+܁X`Eϕ{+R}VsF)aC|њs&h9G͔b(\Qe8m,ba+Lj~P p{T9e~s mD)RE#fr57.L57M|du^:lӇ7^C1շ@>;`BTFR 6JT 9q=$Z12XDLC9Xq}no tBhG8:Qa@%6͘DQJX#60Q#: XV.e0)-YiVr^H a0 ¯RnF}QyALqJ>_j g!0.a24 Q)HN5 qj1 Fb9^yj68flΘa$l)-xx.) V咄g 2gkE6YEŒ@KQ i<f  p< g:VMZ3Nή2g5ha$-]dP@d *惖":5Ssb4_+9dIV !W,: N 0C`]+Lo4T?>7SsU\<($8|T#@ӶJcL2d%S;KkW ~ mhX(,}ηvy٠:^.} Ub~ayo\aZ TSL=3EA'抳N{V83 ޝkfv[5LފxCL]x:)EQm(Zuxk$CCzkF 0FcvfP1z>Pu*p;5K*_XԼlQðݚ T;}E~6xDҼeLSj/\hBA=IxHJFySꂛ2a]Q,2" Hx!]d kGmLYgzk1Q(a$DF,7N;?H+P!Tvob {Mkq_kg$U }'d%*5!l_> uYNn `/Lz|VVJ:* ѺK}ןOS,W-䌲ҲcLВM=j34֑7` gL% t'KcQ{ ~#Pg hT/X p/vtg1`=fޜhdt}$kCz^w[0s62~"ǭ at\%O)H||_Pqь`9KrN9-}p7`CC|ZPDuj,]^ԚJ5?BAtK)/g ӤGWQZ/VT#$ξ͐Tt:7WԹ% ֨m/pt; }XYT!DG^<|_495_m'ڈC`:Įdt GE45whYAЋ2嗱qpwgNˇV KZ7>gѬp3L8V`2+ ^݃Z뽂dkp{Jw'BYMȥ\F} xyUuڽmMU& &yx kn4 zL,8-^l=AWjZ(QuH9&gZ3iI4.YaAb)C˼a1(HV" h#B"{mzvm& =瀂kyջ81!X"x")GBAx, b i`!a{둵mDZS^GH [ AA23^.fQ('$.Fc #K݂Zjߡ{Ҹ5Ku圀^ͨχ9*=Xk臋e,|( wrYo?|^յv,dZώ}tԠJ4.%W"F%.*LQ;+3fl)KJK!˘".@H6wDM!kޖ֜ sT Yj)<)Ͻe_T1XNRBݲ/danՠXiI`-D K o=esWOJy)Rù]spT {e>\:Iy^ [/`$Qv*Q3gzOF^M_yPnP?p0⒍lյ yO Ӕyy̻?1OsFI.S?RFJ$zA[2Vne2?~۩TD1:eˠR tRy#muF#t]!Q0ͷ캁i9o]'L".V~mEJzk#KQ#~1W J0/4i@ Ҫk{\~e˜}N^)'ôFj{t~2 wr_bۺ-Q~5w|8Kck;yJ%U/4&X/#Ҝ&X&5FBy>7oXGu9(/Ŗ73Έ-7qkk}bpnm_}niVE{wvXMFWy^!>T &&hnXiR14L%hg#z+8 \";9j>ZpGiëVZHŘI58W5,T)q-NIzSK g+ Q5K\,m-%pÜ&RJɣ D3j4l#QR=@H4V6PLp c<E)AOP# n%݂j!xxx*ؔE?aΧIK6zWnSM1}3jgz8_!'gww?챝v7Ad6YTH*v~4=W.a[pꪯoۥL-(rD!gqjxms]!#E=ځU|LysFRpzK y*mn,ïo%jomI>Nf&wҝ}]5/eT-T@]~_HR*^XcI9U`p!eL:%B2/Z )8 ڛc{Ɍﭙϧ42~6]w5ѥn*cP00Pq`^H>fe|?n "y-߮W~Q$&2K 1~E(1"WdOYi/G;5e1Mm31A@{xF xDSfK*N>Za"}q ^%{;'tn`=JqxwjyУ1aԡ%s$6a#,{/nr ]JUdQ1Ox}U .B1a1Ch#frOa2AfcU)ﳑ1.0 %bJd~ p'Łg|{T߀wjANcQ"- EeH:7 ם1UlkJJ֔Y芽Fb N5D@C ^T|/u9V4\C*pX$ 4FMhK^Ge FܽC'(#6Ȱ&$]j貉ل h - Ÿw18e衟 2$@IlGGQg(: )<Apng\=gڅK`~Y''[ҵr,'ӫ?-تm=6Orom.4frF{% ̒Ck6il?_m Ȣ!`D #TlB! -Zܿ['ݷpx-*yؓPZ f}td3r8$ltOGEFICJs5P GSt:ȫ'm8g/s(X.G [tXK oQ%ŇP?Oq"-֖fI/uRo?)UMJ#- ѓ˩?/>D,Rğ^"ыT$zD/E;tVF@9eQ:h i}8'hR׀p:R `42-~bf>WnԀz1I9қ$p=9ܷf%ˁ;<ŁbnS"'s!-M( LkŹ10oh,;i:A4w[l2tcU  ":Q* T*Eir+ˉ69 ;"zBF8-MJ!{Dw8dL Dwgm"dl:o ] aXd^7lPxLJqowHIR3{yRwnrt-$7P&[m=O,UY^g}>*}V:Ci5]yǨri-{c??ّn[: qbuC eť,:.<&EKK;+BI2qk x"2upc_sF6p+ 嫵WXepa"JU6^5v0EAdžfoxL®oA]B9aEeK!R`f8_cAhI! &AK2\Y٤2{s}y:=~kfwUb?ۯJ /}I^kE5gY rnN㋎.?^߹ԋ7kWSsplQ\HTwߞ||g,A$2f=7I)2gWo{/Bt*Sr=ЃfT%F^_>;~9@xƨ!ryV}yfB:`3xcsn_]1ke1f]5ШּZj5Luk 0K[]~ZP[*]l)Rc/]=ױ\⫼:9Ah ͝tv~au /FWU\;00:/k #K[]?*$)&Ep Jue4rTtiVw Z{Ä 8<G)p,GLXK4V3L;BԎ ͨZYkc!T$ka`aTǛ-L1A`(@LZi0徍KN8 W-Q#{ j)W92f,_ܚE Z aǢ!,IUAFy'=5UQcntc $^(+&'JFHYPK㯭 8U>]νtbЂ)$oWi,mxƃQr o!"/9Gϧ7,:(2AOF_bVdOLM]*ͩ8VQ3B]ŨxRT6vOyj_..~UւcQ`S}YRdXܔYoS Ț?zk{w֚3a ЄDSm]ً'K&^U&&b6ipCf)_ǖX1 P*'sdQ`)cDV'?ThSv1C^l' Spk) ɰGm/BPY⧰}>ϿPKưxakHfGjjlR+evESd:ʠvQ5fޝ}q""WӒ\/>DU|BҤ-f!LG\/޺OTiX[W^ k_(dMAPi-zAxGFC[Xcڄ(spb; IՈZґy )APԇ ,f+]aZ [Fk?>2p]r̀r5`'ȟ>1C.=M^i M&(x C҉|l%(PM s0[-@v}0۪|l7 c,!O(Ls [/!xO{xe#%PaIk U+,DBؔr:Rr]ΘGSrׅaGl)' `ˬԀQ qB1FVf)32ɹ2wUh(q2ނ@R6Wcx+D*nϻ@(I:㑑C*^P@E9F s})ށ㞛2H];h񊠆yA՞REzIj)e e3o`7T7p\D=X,[+ɻf+k-Lԙ#oܑ)]ՆmD]2w 6*\Je!1i0b3!UFuR;2bJ9~ eʩ9N;^p&6'b]#u:?lxUn FDP,`‰ۋ`7g_[vjrv]~XIK\)!<ԻZ[a#oՠ?Tot0(%k(*Tl ^h"^Q*O3Z(Z'kDi*v+4~U$R,Q֦v*AFhas-a;R,9ǒA) q="DfptĹHj)HgC TZdw}"V/o8>u۹qPMNoSvXfUʎ.r؆jiYs.ZE;CUB-m[BGqZ oUPu պ暰S?0Z&2K\}8Rɷm^M/T*e T FSm-Q|3wZZcv #{EJ"pZ)vŰE0oc̐%=j=K**%ƈzJM˟{{ Q%/i~v)^>f ܪyxG)en^e(^Eo0N3~n;nAքaܺW[z RYy}ⒷkAʾE 7yn~zpQL.G["tV{t9T+bmbwv7p[p t|075ʅt~ՉS-9y*2*kp$Vox4Jﮆ5ړ5ʂ+9#VnxtpnahJ Zm3`sK“6f6 ?O~??D dφ A{SͶup&E4Ӓ+LG1թVpv쀝j1(i#V;FNwKJ<tWջ]EF%L]n"#r@VAԎFn<Ԙ[}uBօKbBٕխbNVHz@GVdRŊڰUl Ȇe#P^ky q`\~er>q)6MُqZ xձل,0sBA'wYtc)AźkH92+o7P834e!|:qTay"plRJ䁺W%x]A%Y@?TX{S_*7{Nߧk?MQWJ5_oEVFk,bTE؏ڜWB:eTSF+%^ř7Z?10\&Ui*{~p f׿AZ^m JHx"Lȍa'״4Wxv,Xjsn4&‹0IìU9+rvl=[%/̽mje7{Gz~Epo(1ZA_c~1T1Fp;[u͆Ŀmx! +q3Uf[ʦpn̞eb*D_+SS$з%ZQBUO 0+@K}-:o+$RIczՀ:8 Ri&]7’QNѣ'Ag]4b#%ʣ甤H͵U>DlΡB-%̴ kO}6[%@oc̐kbTP﹏Խ.;콹Zdr,\pis]57iQ~X{6c]絲[p d\88}9 %%λ o .᪠WX^1q{DAl;p+; ؛fh-h˓kJ6&W޹%_ϲkO`r1ԍqd !Di@kjZw3oWvՊ.mأ󃣊n}iތ,(%^칰/ÇQewV=?wV˗_[FEReZ2"ª ޛSL[ VVf2TAԃ}kuHzVMS7ߎ w LW:nuP^ehVh~A-4*y3H332r`57&_--)kdtuY͉;,S{r3>v *GŮ,ў/t&>*.adNoFfz&ן]z)W[$ pf_n*AFMtjb^Eqy[N4`ʶcJx˥dp*O̖1Sj*\NLnד\˛~\l@Z^c#P9oC!Uܠ'A[ 8 CPu׏ 0UCs ѱUe{VA @PT|^(C+,>p@FeE yO)Z85-wWJRʷ|X8\DP4무POЁE."$'y8 TfKJQkLE K&h#iş,e(^Te~n+ VPcYN3ǙSe̘RT+0^[m1獶 7R֗kT4J*JFAx `]+~\9i諝^\ή?'qX>xlL]7go_lgЯ.*^]3yv3ջ&_&/ ݶ߾wn}'+ )cY!KRnORH)]69V-4{/gMlmHRw}\Ս*͇i}I^DFXӉ`Xw(eQ;GtA]!OΌ Rr oUYĊC ?C]g峼{9E 2ƑCAP6c[|4 #ߦE]۾o"7_ >P6΋\s6P}(uߪӁnt/y//oF)1o=Y80WND9pLAN@D!iLU‡D 0SsP.e ֈAdk <6r=Z.c$5.) p[+nb::׍rʆkP2/ܟXI&}cdlX#9HB!e"WR) jGC]*R ȼ_.gq$lܽFƧ̿\z~uB>fo_E|YjΔjMO֝/|ww_~=gtϯ/x7 Ov)OWo_1`B$O=N~;X%Nꠑ&NNy3gjolxY%HhCՄh́?|? /Ɲ}W{ߩ67kjahqR]Y]Iu%oؖTSΪ,ߖ 1DI%a ʛoCIYSj$į+nGU&yfjPǔj8t΁6A6%W9 bPzghh*"mwmJ~e/) 3&M6H'"}ݲԺlb*օdU!*X5Yu2<.l[&Z,&*YBg8u$Q_U:̏|a~;2;f<<\a)6̓/5lÚYp;Va9 0_~Pl $;/0_^xVHΗi'_[u(ye3J+x[$ a-jMNEՄɫ؃M]D Zh~0*ƗCU,4JsiHgYq-sJCRh($gcA iaÙB +V*)"s3=8j@Pt[6 bLοE토|>A2a*wE9p6Q'1UMk%[A[m?|Y=Eԧ),?u>hHznys~lЊyXͰa wv  5+`Q^)7OgnjU5/z =xQ'$t AR1Q|(J%T:+kT)Dk4^X'-V>[!P;! 3J&2!倝! Y=ԭ"r1\p,MtXv*b9UD}JĖ*/z8]z$Rk*ZrqǠGh䐈1$02'7/JZ(*B,}^Z@"(:%>o_@^3z8Xah4pM4qeԿP2%An! {ik1DC`$`3Ȥ̒kCu_ K,3+}TqSLc>;4`sI ,8qY {cx!˘΄gCN{#RF"g1`c[a?lR-ŵ[Ewk$31Gk fbh@tl6hJ*/;Cm(Yʰj j`6 6TCm8\zO?N"{\~{2G_X ||_]>v}\fb!K!J 0QO>OcWo  F=Df?&Q?n0Jq$E#VE)嶶{>] \j1R}.k N]C^7YvhHUU}*C%@lӮP ZͶt$TzXl d7t3pooJ 80;e1"u'KI7 5l_}cM-khjQlOLÈi}DZJ.Gu῏c`j`y^6$N{guĸk S7*ۃg%y4i1| {3dwPC|gxGYll>Dk'w_=|?}MͿ~ -J Zv{zsn^}nAHZ_['x1{CL*&4ms /h/v?9wqKT(ʕ՚t'_[$t"%gJ £s(Te0L\OC(+zׂԞYcVhtUH'O pPm:l(}2{{ގG;7|;h)_΁{oRÍء)DDjXʑuR*sf\]@ ]ЁX T&R6Rө~)Ud]YknW[3M@0њzb.d4T֎a0$DY^TR.Du+&'S &f・[wG@"Kp*M6i"HJ>s0ZJߧ>T߬o"  BE>ypfї}=Ћ m%c(`LuwT+f=#O$?b9u.L*:tu0=_"_<"' nOEPgnM(cU2*gPsLyTvJgMDiJ {_UR3ZrK*JL8d'TwCT2AT1M ^Q1n×H`p;S7Bu(\y-C6 D1a;zt֊bT/nDD!Fx_HvŌ;EPF:eGd? ){Sp71,V:I5lfKx]!S>={; 9Yp1dnPxp*Tleh_nnvaea }[ ډLOڔIeZt5}~TQXyG!'Q]btE8OQ.jMU(RKUvX*֍=Be~W_޹H/ڨs*5#=Ӻo]uεyL?M793&gaǸ:'GvϬ{$A0cayR)ȍ0tZ I;@YC[ۓmu7:=m E0k:b R\~!2*0iZY$lPb[*`.UXfJqir^bcRqE-wv`6:?{W6j\ xq?lloUV|+J:WӠ4#JF RԮS"O7 ;\R ND1[j=%6Y<?gUKKԶ $:IJB:l_'IK젇j.(-eKMmz&^[d?m=} wJ^(*Vgq5}3Gc?<]f%ߓ>h\o0 3n\DDhѤ3f?Ŭ/kcz>E3Cu΍Y,_*~~?A6٠E'%o=}g%tYMCWW?Dǖ*o*2Bip;H)ۊ| #%!cs1#n)=hzEf{FN# סkH986(=[-ƚy +$D)"9 J(gФrjy)*$p%ax##ECq r/ Pѐl̒aBpPN fQ<"(DtFZ)W(P‘C㑲l@R~h'!77]a<(~>Q Q;2;ND*ּ*!Hq͝*aR-*Ikʋ=&zI0&Lz?i) :[C\׵*/va?^/jxqk'DӺ,aUVJxYWՌߢLX{FMk[l/j3 ;WCBJFaQd˂ZiscVK{4wЋV*9D\zӮv-P 6\vh#c1jdq] Si] pme& Otk{pc)'hcl(.-SdTX)>}-w@*g3#sO&dDUlP຺,EO kL0il ն^^Tjk[ɰ"k^w©KXpK #1-@%G-{O]jA0EH+2/9HE|m7 (Bˇ{ⳋP :}LTy.%0K`blmi&ghDoZ$ctϚ'^hTIW F(J!%0fK[*L7{~u \e]YS#PvV:dI#2+$-U5˲D*dVHO0F0xS!H>C@ Ч } bS^t)˽ZA T!}>Pe/"OyH{UD< h$ qL& % FStSEH&q/5) &;mB'QTs?:<~@e?PE!RI[׿qIK%i=1^\2?f:t& Ay=0;LZNQJRs#rF;c50)v2`=И*O[;tOiJO Қ X6#çy87wc@sx\'kuA\:h4F*~{sȡ:PH)ŴTثLe.BIuZ.]w-2BS)pw{P{= VmB'h賰0( :KCb"0:]td}PLHQ!(.1RntW|B)ؓEg"[B}0ߑ`n͝JSf=~Xk =j\뻓|}woM [* . =?}wuߘr/99 b$F:'~*Ωdlih~QUȬ.),I~Seh=ee($ND)aH~?[QGi%(i,C&䤾F$&āZO?n) giEit&tKEY|yK6LB|^qU}fFW67Tt+%aQ}^+ ''IJŰW;azlY,kHW[VlQO=u!J) ($q^4Mhd9ܮr vϵ(Nwýjv/%M{~Ly96I%aro~_gLY |Ea]0=gsZ!ϼ$<Z}7mP6-~@nG-kMP-lkE\>(|—#`0m"R 3-FhAKcoHA%[ '6Y0mKc,f?6v fE@Hj|vz/0?/f/RO)ڧ`k}jv~}_P㩩x)1|QV2eX#]Jh] kU\|7_`6l^>'b` G.O+Ӳ~lywk]5CGke FڛBt֑E>*[7Sw6-BkIk}O+JGfK?R\!GFpWV%Tz4XEҔ\ʁ>FBW|NzIۧ{KMCfl#2RIrB3v6~Oj67k&MupTdg fU^`?[ jc>&cntV#h7R!D|n~ zb9:U8c݁WwM}ꎛ> z|ˆ*Wf?)͈DE_~[vY^9s袺A.Pn] c:=HLi #1w_(qkcVbdbQt bӳC`O1uQDIטJE[Riwz!{Mfu@i},;w,V{0˫7ѯqȫtd2 #rGnHLfԠ{Hp8B'B"6&֚^o9":yD Es\L6N@q}U\0MssS w aM'tEJ\(Z.\IL&aE2i[ULr;28+N5;q+ cx3mG *6`byq& S6`ih;ɟ;,"E}S\v^K*ۛr)7<yh!$dw,4 }' R܃_?D9[NS=TwsmwJ&U^Y*R̓ ? _:x}zk;.P_omG]¹Gg:4tbzDjks<Tm(u_K"'r24G7"ԍiL; [?[\]=m1~ޢRH10}Xͱ6dyB8FukMK<ReIy& WN{m #{m=NXFl3}RY~S ֣r{w`} y TY]0Ek3aH;lkNN`mu1_Ye^U PɰTL4϶Jٶ(xe&dJsU^[^Y+_ C0p#YY/l ~q6/g* $8hnK_ጵp!v%hUV )/54) ]0 3.(FTBsAF:% R2,ʀ.V`Fh*mxcƒ;E*m 9A%C`SK+WZ sK%B^DCHd]@1&1,< c[kqOȋՊUz19NT VoZJKt9eZ v}3/(|w~9A'l$&R^$1tϟ_43& +`7E4|`[ĂX ޿uHWנ&5訋ă;hTbAH ]FzT^R/+ _vH(ay+C `n8T0?%fXD|fJd&(w`_$ʤ_N+z5s >bMq*}Ӹ-WKؿ9+Zxѣh" E1&ޢF9IpsX$Mt}"Chǐu䤙;Qh IsUaDЌ  /5LJEƴƃp 4{$IIRARq-pT# v]|x>\j ؄-3G7& L#X1VcdCyQ"GVzD-:*]ZUQDITђ{PHo=i"x[@/;m36&bؘWe W_|Ϙ"8Uj}QطOqFFDYm\3`勴I,ExDWO^슺E6NK2k[H=p MЌ 3:vu]oFZɗ;ߕ `/j‘ubyi Nk:?;.h{b { !'pp=yAV/>B5̬ܛi_1o5u WеdZ=0VWp}jie#4s'д qW3K[Yri}rm#* ]VB֖Y4`*u?ݿZ O-rb0GU 04BfĤsAf0C6Fu>*R-GL܏=H\lg@N"S dR'+iEkȁtvizunz` YMhAHhHYEץP!d2܇ iNսsALJ%^0h"Yu&>mD j5 U bQlsR%-ά%DouIyՆ3ȴS;ץJ\r`6eR" 6X؂S(͵-z'5ݺTsՖS*n?Z_fi^FRFG_¯ ނۃx7!BE~ˏu%*J?|xOO69o+L oo/y<맻W)VMg?,6:_D[aF=|GH PXa._,|(n4͙n-^J{Wȫ%Aʐ$HIĀ,ģu$,'HaJ6 ؁?AkGf2JKpelIi aR`XRhGj_r2 S͂`9Z+aM+pJ碟_R/ DBh2G[ 6M~nry=G_l2-&ohu~rMFh?uNj"˿;^5=Q&;CʤqnUjEZwiN)^|l!$`nvI/ ĖvkSs菶6̤~Ӻ H HvRbssFDH(ьd@>twip[ƛ⻆qhنp>:4 +\_fi]Vr3wzZ)j:"qK߬]Eo=1k5QeK{TW^f(cQƦ9 '\/pN QB=q00lCimyY+^2h+S۝in6S` {TR`:8B!9q Oj|yi@ꯅ4qcO[0B1 c&ZjgF0  s|+_Fpoyio+hS'9<6( }B(KɌ8̮_ 9mNrdj9}%t=vhɴ+%O`ĆVq52Gn>l?㊾adwT-Ds~QR8 ߶o{Ҭ]Jy#_D5F-Ȗgr ;T AZgZ3p锧0?m(Tjz6 ͔=fe X+v%h R9mvҐ &ųP0)z.M_(f kP_H kzs^ ž_2 1xsrlI=@ԕb0 !,Jԍt*Y댒![8CLR8au2p2+߰+ WvKmfhMH/7-2Oi8~ 8 )4F(qfOUi2¶j7;Lr%sԬ!So2Bɓ:RՖ1Ec42Y-z NH\C,x2 'RbsnAږ Z;¶dQ)LD3beR\ƠN(UbL{YjZゲ_r0nhء<.e05):dI0E0I;&L FkT8=2-h!NXJD\Bb@( g3M[QdIԜ% $v"Rg!% '@~&8Ѿ@* hHb'tɀ/X ;atN>XUu@jsfN:dH yqXБb=r +nJ{e7GTj۩z\Ev.YJmw#0=|\OhU`FAiVi A>%H[MGx7 EElCD$cd2Ŭdk]CIzuʐW9cJU^q y$j,I;b{ɏ%ͧ7TgL"ihj&uKj Lyc, &(1|m»qPˌ!>[Iɚ1d#PܱD]hh^B))٧ߟ*؂Pc$;!|Z Z?׃\..H,{2hРBUBuAU4]]r&+7w;zi6xQ7O\^r6\7po_|TrR41"y4 =p_xVibLk6Z >ixn!Dp ZJ-:!w #֊9u9r1\Z^x V%و3G.0J52gX`x T[fBV&VRxW0% w42W% `.Xs|@}مOc4'!>NSYE1.8yٔZBa^mZ鍟 ?Mf|p[x"!rguihYCk.|~icj'yTVM?BcוESYI9"sp~u5l'j`6)r[sosew`GN'};i$\:vun؉B+N9`7Zª'*)DqK2 .Hd6t%/WٺF&5G"99`V<6VT餫yg ' VaST[Yߪ̈ڷZ̓8ev{B:^koWv({ltO -l.ʈ+W{ٖ<&'#y`JWhkl:pm"0l֒.8Ӵe{  WW@^w櫢/x.vZ njm7u,^7Q];Jsjha'h7 UvD/QjP)8~*eǂy5[Q\NpA igqOƚ<'Dɇ )u1Hgoobbm @M\Z?oz,߶P:^=(kFy&ۀ6 LI(D97:HhPKTBhB[0(`J) 0T3i26E:ä2.gVhCɲMhbd8 1ndԃ|t[VR+Ιa6'e)V߿˻osl\?Mk:D+ǟӇf,V,W7t@oHcA-Y7T>O:aAg%"և\0o\^jty"Ө XRmDq#3DR;IJ㜑dVSٕdm`l,i'LUf^UOի8!iX3}Vj7>[TZAdT N?\)ݴ?7UTcDzS S}%ף}8fJ7:\yal" E4J9LE{S-h]'Dق<aW̜-Hzi{ۏdS!XiGCX~䷃eah6{X%-x$gܜ׿6χ&~CW_9&.W!?:ߕa㼸u^k.ԕ`4~L@\~=|8ӵ-?^@SIBMژP+c ' ̝ e٥xZ[ i5a=ll6߁sیv $1+ kqSਖ$[):W^hʊ)IJ]MYnQyM]ySk3΀2#|0*owRQUdZ Ɓ@죉$?A,=hG,!dm8=\O_ݎWˡ_ Z}8oT+|f 8 eRbCb)lE+`M]mzd7t`i?8=lʙ8I"ϣ8.)WW*vsxbG `GVJpPK^pe Q@LcIOҍ:3Nڈ!ENT;()i '@v\ ;ے+qnrH*8/Q-'Q92Y`YKD7giE AHKÙ6 "Z"8 5 ,[JtU2) 'auQN,*g4/ RΨVԟ.ll%l[0´nRYOO`+U(=͋<`~ج釋).77*7 Ro"}Y..{٤,9g o_ƿd7tXM_{1/,H e"#%6FNinpѐWPZ9!nPJ\.fsBX&E'MoJkwMzo+{S?kXoJUcꐈљ4}iIh y!-X˗#-ѠH0JZi15HRPM,K<9j .H^"qVVFA/2K ?en \c*BYJKJV~|w`25wm(h}I(㋝GF=}vH$BGFLi0.^5A2mp`ƘP(\!`1j%+t^ 7ԦmM A .U~Ҷ0 !C@ BSʵįL&Ako)?q%ւ ǑV?yI`[!ۼ3X]OBC"s"$HSjBHak}Ά*AQ}aAHvĶR`\1F8&"^NAaW͆@cXKcčn F~jl|jAY%XP|Dxr|h!۷M}BD7P<`Mپr`7^XkvA6-Cc,w˲= rpgtt^~2W4nE`zH\K u0!2;[{IC+HgN;ޓv_-(ph0Ű)"]|?S2&ٶ_~8s[4G!`˧o&%R`VcwJ - t*\rn<>ۇѸ|nj/To>1"l'޹ܒVF#nplvr޹Ls}B%1GTkEd{g1hZ1Yr ڴvjEBZd}|piYWi97ƾ1SqT> u8п)"TG#ri0/.:TS ;G{hyJZ Ny"JOE"*w1aP铍iUKsjB\ 0Wj_)rO48ULM(1ft(nj#,i=Ud37Ln2ZG\xA7"LUX##iiP+-1 +g, b 9sKGkهFߐZ|c. UC~cB@^9O8}ߤ}JnOVb<L^Sqb-!~z2.-Z /RwZaFw4 kM~|?f)V">-Ȣ~wc~R0zXJ =̤;Dit:sK ,%︹_0 \c7?Z;T~l*fP9g5RGcw :'5뎒i_՜hT:ow!Q:i#,oVvsg {Q_8{RM oxSDs$}UşaEnf`Yt^Ϭ2ol%'.zAYo(fp I֙!MD;8mFpʐT=>[rEG~&:LxZʎ>P XWP61J]& /ކtĉɟm *Xg ٍP,,5v6iP=ot@l,0CA>;3Za|Zj Uֳimf)^2TytyfA2~+ѧqc]JJae뛛(Cck`)BtyfIѤD7nZ%>qo}DYLqX  Y!bKZn^XZS!4,shؼT2a2x!U܉\(e3ɨquQg!oAu,y ϷOWf!2xLPn3/E{&3&=[pLvԂS%6W +HY˶攉S9d#,8By1ׄHrpM/u%Hbnc,os؄T+=OuU`! CۑqO/fsWlz6=:o0$+W s"w9/Y^Xnc,xYb~fB_ NE[\Va;_~ |<[n1?1Wf0уǻM^ÉQD=  %FM(Xu,iL^`ƅ-8wOnH_Qӥv1usUwg9Ip^4N|5)iDIErb"n @A Jha -#"Et&qC鐡?AQ1-T(*4֑U"Aa-R U /ꪊi4onlnp,<+Gx+#mz0\[  &1$ 1qV$dփ$N%A.y8#;h_~Zfg5C_};0"8FýWw"l]ތѻ^93kRSϳBc3.g#~_`|"gۄqb\uA5226b(s4\,ֲ&]<9~yl,c ߾_u {@8gnpË Xp"xр*(H"0,RPG%&&,w,bãJՋ Fs+9ԃRh7`SG I> 6 K4 f\!KAQlQI* `"\al@:WZs O&8tsB&aK\J@as)XdH5S*S qָBUbBsk'B Jqn*cy9p0,1Bq0p~MfʹgA*d2[{9TX̱D$;U< T?S XZOZWBfVQllQ,kop6iй21+ie_S?ScYۼ# 22WsJM| Ic 7d'{Aљb/`eQo1` ֠1(|& WJ+C~/ϻ;Bp|)YeLl3*azt/~^?̜?|3`L1H_o~J)ʙx5ٿ&3 6OO rg9# !(z/hHKRTdGfz?d!`4∁>]y2g/V<6il΄8x $ĜB$XЕP(Җ9L0 'Y߁ (2d4G%-O紅z1P/pfSJA! ilG's C! !Ac1&Qpݝcqq&Iv [%+'6qRmV ~]Z /_?k k&>Y33)[A_#v'}3C vyyճBlZopL"JRTOډCı}F~Cxȼ2^zNpKnq;WSTVi >"|ЈxpyņŸC-<#6ݽ泞 Q \nָ BMقB UأߢS\ㆆs8AIv6 %PB2ҡ#SAHzuGKM;;xwO@̘ZYVi!a;<(~6BHO1ׅ!Tl#3I+=L`U۱i}rzJHohoOٍ+e5EA^ a%h*gKwe$%xG{B&{tt }G* uaN{i x#_-`]\zzI2=r<})tj]VjjBE-/-%';)D:ɚk|J ovF+6qnq  zO.A?3BOxba W};Zg M[xb[hw[a!x0#]U;V/Nw鸀RՌ`ꛩ5Yُf'DtmM)Gb2Ds ܹ h߱|[7na\?n$N)cZY~}Q)R}4'w6Ȗc/9Y iYh1ڏ}ΗtL0kDTF$DI bVK|ĈN+Bj7WdgyyzMWpa6.-nyP !l|yG_4mQyB՗n&yGm((DS%mTı)09}5v@ԤgWn] ĵ .nJX JY:+7V VE sЉ✄APi3abd%1JHCV%/[XKuuİ;i͸Sӈr+U+p[N$Z\s]֒V }6\ 2W{/cvO{ōp !#> !v? 꼈'wut@iڊA*ux.`]ݷζ޹\ jܙǰ੝{ 0Q_k 8'/雝݊|o_?x>{Iu m;U?ysI UE>aDc31O+C0zq?,(xf`Sͩj Y|Fm~rX¯}`^,OqmrۗbXϳ`ZhT.O\pe˔/*'D165嚴\|=ȭpDHXlKim@J\8)ª Omw5 /}]X z~.1%E+D{.(X"E<‰W}Ā{+J^w)1vT̯ĎۖW`Eyg՜WVBluWA UaD-:g]Wq U;'wd N =y\ "ϟҮk?yJ]$+aw?ł$iJhS!}+qsG 5ޥ(@Y7݅&h+FayVnv yA>f8s1e:q/M[Qo݉ˎ9u=H+ PW3 ûyֻiVwJ2ĚXXԊ(K B,D!$DDž`Ŋ$4wOihF}~HRtVv}Q%"wH"hZ6!r+=DʚV!MZ<\^u8Wfp[~^"k'+̯FLYo:}>/K WJC[;Nws{HO;g{)>ip4ѨE0itUsuJ#z?d;~a  HwB bY@I̱q2\;܆֋9ALu;z.N|3jޭX˅c"Iaa֊*7ofxdu`9 &֚c,SEZo?w; f]Lj_$\b_'5{nwu3@D:8p9.J.a>uSuj9p818A6<ڃ2م01Jew wE uw5|pV@3E4Κ_i xk85_=$%stZ=F'/H]f1UqɯjjxWeW`'Aw {/ҨkʮxԨ;|xli'u4p\D`;uSM+qKKVˈֹܢ:U(a5ϜT l\}ÜDT0#"E4 $NxH#28QcZ0 QU$i#EA߰JP1?_wmsmf>ƪ70ET{M?Xʺѭ;d4Gҹ6>Dyβ-+I5w>[n]w]/EWlX JlW(X:S@w6N?zAZ^ WxNvpP#lNH;@Szp֍ CWK+ZcFh:ryt i(9Fg׶.`sټ1ԱiRVǴoO 6{?ͦPvy'Z %T,kvA )s$K!Ŀhp\,mF*)K[ scZwIHjz˂kW\aJ/Oi,Bͦhk4SY?V`hUHJxsS DqDؽb"gtۿh]gR|"Ho_MpPnN].-ĸԘnMx\635xS8J(Usvn(WC9¡H LJn/e%_n/~~o1}GNW.zr>YRJKy?yskdiIl^S1%976p PG!)+mOϣ?tf)¦,Y~{Q.zHeE?Gٷ4/+Yo2-|zz0B16Tf";YNݽ|f o{=0"՞:uSChsFlP˰H6՛ّa*0 8b0=?{۶쿊?i{L{>vZu SM`싶neI$n~gIYd٦DJODp'1fz{!:'C(Sc7vC=T؇vߚ=}W ǡ#Lvۘ{,>OO>:}oסP:Y׸3 Xsv"\^_/| ?e}АCpp*\Wdi?݃EQ@MN M'At<7N:ipW[|{^K 4_AG7B9G/UqW_@jB@3:*Ys]t@ˡaF_O?^ ;؞_>Dbalêkșr T#4]Q(a4ELbB4.gRoAp_ڙ|8/򵯻oF? 0OsMgA.]8Ɖʍ-h<(J*_ A1/Ȥ|E":>ț>Uzt)&cPL4ux|t zC.0ñFp(+ ,Mk50&ߌ']˛PtWބQdY]G`IɹA F$iM=Tg5ٹ\:yhY,<: |59J6dj/`Ei= (24֪cY$.O"baTWe>P h<=a5p,k&{Q4e+9kPMEB|_eHUtXB nAt>wk/?|{oUb+o6HrCN̪*gZu @e] zUL4raQ=> 7$RR9T .A5h&Ie^6UJI:tZK,\r=%TxK=2 \~ϥof^c2a\a"6 I}gmExլUeƒBH>7{4a_STv\D:Ibu buBwRE*1ok9άz%,,X9 bs*rtύ)<SVٷ#0U5lhˉ:˓4 ģ`54# 3/5r 遉T2갤Fi@bj= ITX$jRV ˇ n{ghffӊ~xœ$XB%L4FTxўW]ZbDf2]h@u6[op9a/&]<@ې.dH|ٟJ*~Ivn߮.4'.ؚX jaN᛺NO,ɬx kk)OAI_LU%gT=q%O21l9I_tgzM6uww/B&́mF]en(/{zl\ S7\("?4 ބѻ⃙ *W5VUTW\ [Gոa*%F M!qM5m`sݸߋ`Y*ؽU}Z!2#6/*d m`KkPԘv$t7!XN9:gV8A!OnzXL񀁙)+!4Sk?-8.AVlmh{8Vh`ua@׵P8_׺ ʵn7oغ2]l@W?=-k4xC/lnӦ3rk۔\iM)_}ojs]Vs2m5BiuLbWC[]V6[k%n֭X E,ȚȦw_9Bhn=WAcUgf74M# cOӼrx쯨}FT=G+)\q⹕o3V;ɓsREVǬ6 X1Ga< :(N1wFT &htg-#ΦQDR x",^x]@MBflX{TK"Q STi/͙΀3(!ʖ3͜O4`Χ!Cl4IYJCR[z[%&*S,|gVPšt³"L.">AУn0QƃV8?G'~rcwd ?{gSJEЛ| J M ^̌lrZ^DN·/8N{Zc&Qc&Qc&Qc&eu4=u[jq,ީcvրLaʛ^tDO)RܖIS^Mߢzv ܜ2~~MU'!: PA Jcf3ˠS0L4e"`pJ *nƸ͢Ue Z^/ցyμǒ/L;^h伶rKaJAk Z*UcoNļVUNAN:J.0t r/()V,:7`OMǀ։֩߮;hx5 QC6u,F<GS?eۭ5Y \|i8q\,vz--P5^1').b/XTTowFG}Pg<r) t+B(vW[A*2U%(rͽ`1Tr ""i6<ՂK`. @wr44EFh(/_L:"x\t/YYuPB!j2&&$Jie(5ۿNNM*J)ٓq~k?A+ F( n&(e8EHa NQ Yl୬Ƌ- gd~qDOxMm8;Y4Mkl =G / i0=((`kХZddVq Xpdb枻: M,\ѫ:9IKN9u Xd F8a|G7r{ *{98A$&?wouhŠWEbupm!lsLqGЖAqӭ˃AD;.VoEcŇA" -L9b1aHm%syJ%D1ΩJoM6^-E -ަwc13֤w3z$g?)C@jE7G-DE(QeCv(^e.V6IfS_XnZX0tNF_enV(4)r6mLChA7<[Oh@ 8.'Jq "yNn-{ϙv~{{P{?-x 0ӍARi ɵ6:/q!ʝbl`2Jd,S,UԆe X%-X.yZ zP)!nV#*PLvTxS,|P{A{/o$WY1䲠VOk2ZKi^ e)X}>'541$:K+XO~Ճˊd=ƛ;1v%lc> b-IX4&^!&J2u}ؕ0| Z5=bA\z~%dzgT3*R}t OiϞ43>ť 0!A`|ϝIoMDa\7YAYqnPˤ30M{zxLeRɑeK ĚY0oi)1 Z,nsY)- l 8 ~uqL]X [~ D~-`zGm];p:0qFU v,Bv%rv%96eޏMЫqB6^k5 8[Qyha1ޘ5']/t͇m}Њ IPLuҙ"Tt u `~i}pƥx`s0*{׻l|HS%v; Xbrmjj+{} ӇJV-T]W>2䌬Ou+a7 /Y /Y9\n{(&U95!Ψ y%KB 69#% O=W@ԶM'j38ƎT![Sc J- 8 L*V Ӭ`HۀMB˵1*)ZﳨϢ>5/Gظ0J%X,6Ch:]ֹ 8/{^A@0왁PA/wG?:\2s-:N0tΝV.`l0;Y$2*l-)%骊# GgR$I]2 !CHluPN.ӌ3 7)p8H6 u]5& B^yj]/{[X%4sV ]9xD* w> ar{hcRP9VN``{OZ],@xOM 7!Z 0BhcKAKyG LM0zɃqm|RmInV؀;(DB *&8jeTԊKmV}=Kꗘnz9. [o25H[FpF-!uRw(*FT*.ipU"ԏݯx]-j Q7_9\h'_YhzO *̐_ U9K # LBo@67_fk,%kĽۂ5:-&'pV"5E0A*2X։2IM98/٘#j@RTZKUA;=| >%-)eLZXєsnReŦt?jI9:j:,2h?> }Θ9lMcUMH4i cF]jBUmkf^,8L3 8ӰpHT6#69[W91v8? ~݄ZRSn0z #nn.G(zQ$$\ezRVz5먠;n/r0V98)zo,(~t%qA/+?/~6|1y 4ť1HxY7T7cn\(X_ӟw[U K 7D~)Ꮦ8䉳h"SPut!H8HasԑndzTޤ[ㅦJ68䉳hmtS!H8HasԑnwKJS'crzZ,7T$sxlJ;?JW-B/ t$h#+e@0XN^x[)PɥTy,8w oVd]vud)~eѐHhU/GOk44> }ҏ%>/;_6e #7x]1XncGF7kVcW fqsV = R26>8fJxGo^<̓Lt"KDٺ>;v`KđMԬѕUqߟ&ӴYP.̛̙_]}{g4dz1pם#X ]^;\܂KiٸlU^yKleNK]Sy ǨD$THKBZ(+ڿ qY'dޫ%rHt kk=c&y{k 'R햚]I+)CY-Cs&BgLQ+atFfcXtq,K-2/FBl8]k}kmCI÷;~ϯ/鵹l;g05B:,>_V)c(bxMܓњ_`D7AxxͯlL*S8`Xrɡ\M_A) ({@ %ԞLUx*3WX.O]|&aj{Pg/eIIz$tܛlq1ϜY<~”ݙyv%Cp$=9m©~<{o*/Őiuwio-n4O*Գ3P`TX- K2?\.`8zDxRĺu"T< }>h$`PBHOπ% i>0>`ɶϳMB=EࣇMɨ&H{?Pb\__xbvt{=>/W8ʧ \_͋}E'Ue|3ދalŞ ܭbL?gVSa9͞d!cD>Uv<ZaNN-JGj0M0}Op RtpWYLPEv>j!/u9N#1JMJˏ=O0jJ0b,ϣ  RÞ?,gџo?/Ѩjo7B>3 dc~=gc~=ge fz0q$a6T3`x҆<ƅ'VYmr*cz>z9z{s,|D/G|&:}1l-9Gs_|4O~0cE%iogF;Jg_9e}0>,#=yy@$m1 >[#"ה+H{\2vJuԥRPהp8F>''Iۦp#H!\H+sOb>{prG41E–{X\mbW~94 a,uJJ衄` GcF)l&\4žTf{57=0@Woփr/!o.؇A@e?ez5z.5y$wwoxE#T$훻ǥ輪`WcgǠo󼿽YgO 3_|=M\pl#[- y]֔L XK c];Ѹ%lwh"dL=iYq rT[+b%JP& )M"ҢCSirЫbӵ$b^`FIѽ 4!P<wP:3(E3oN= i5<3}o'STՋWRnP*?MQ+5!ݰ= g潡[|+^Og15UHs*9sDn|- 闎#!bO劔 tQ"gPՇ]:#w}R/{h3 Lwz2{YM(Ugܽ = appt~ͬeJzpﳄ~^9na?S/g]pάPw!i1B!5 ~i/"(7"lfwUB\ӧZvj&׋BIQ ^)³њ .X1w?jWg_`D\x`%H؁D$"jZ\#:`sB f&Ոmm<6s^ #rZZ%s [w؜\kn TK 0޲PĎRbv?"^nbݺT4ʾ}**X~vǩ8wӆɌXɶs7k}}ƚB7~G`tOuy^<콏hMI0wz=?(Bմqғo݆U%ܑELgj(=2}?~8BbzKr*ŚuвIC/ _-j4q(]ƁmT'ard@BjǰaȊ79b̮]K7(_QCrYK&Hɜ2$伮<YqS0t -$G"0cWbg Jʼni'BZ`~~IHT|HkuF%W؞#Rs;k&爲N佳Sg4<ةr"5ƥ&} zBz/nbrM)FZmStqjJ9J(l:׹,B)y-GyBfGL$f|'g[} BYQup 2y vpjHϱ,H tnS8JsK!NP Lѹ=iw+&rĢ.c(rnD7,7oa>˖a8ςg>̧am/Zf27{^'C.팖I}i^!h[neU]gmNjׂ R0&O ֵ?~jdCvx( {-pRe->A93>7)5)l&'[\2HFJ+IU9ODTF\Xⱶ &p]:qc-{z*b,ֆ6cJZ-5̧!D lYr{3iȖ˻l~ M҂vp$p5i1!qւ8`t7aooh`Xxecib6mvz_.@/'68D!@3,ps;+DnLVlWw)ƑJh5/|z##Y1a={ 4qeT,b:%IooCxx-*j@@ٛ_tsu_P_yz71)~6/һ4+齸gឋ.nGZsɷs8S\1c*9qʃ 3FȔR:(RkJs#_NIÁ&1}+4'tU/:| Ds} 3$—m[L"c)ʢ?LV)dU{X  Be_`sA&Hi'vo}Ta)Bm vܲ`Y' c-ZG!/t&9w` 3Μbbq'Vq.9%14Z͒Zs մ4Xۑf)aƷ"7\NSRm-D38:{Qrj0nPPF9&}* SԏbEIz?ǫݬ7o'fɊ>cMy|xs? #٨SE%﷔q%b%my{NZEXumoYғoaUbHB"XCtzCMcޣ&F4 Ӣ Sp:,c3t|5vT`pSe%$&WAP2xs9Z='3NZ**Xj#ccJ0]ˈ8`E㰚9R1DNcЊU 9fJ ŕɕ`6@e4kqFE΢e/ؓMٗlmki%"-ͨ5E+lϨ.~U,Y7L l  H[M`B CSVB~ۺQae nqeeYA.1 BpS# p'j  V [ tҝqo4>-GO L~rliY8?(9bi;xr޹_z:`|!%e&حՋ-XA\ _@bsv+I } rMkC6C N l4F %R(}ĖY$N18m*&Z=ֶ$#QX[ B ՅDQ{y,{xq86AOJx2a!"~.6 *8Yjј=d$gebj{Yšuq,"1F.JdسK'L;qg9CR7x%ݔzҚ*!Ҁ`jAHe e$^ =Y9zv'O 9rr29P3%d!_yǃ9d9ms+ ;ۏL7Œfa{IA/Ef]n1g53^^Ð^cltȴvk1|>5 wc> ϯ1< &v~:xA {|B=ս;Xz=뽇CG`|m*T3@]}tQu?-O§nn=ݏv{]`E¾sR(!TH0I몕e<g6?Ћ<זPky/ݶjv=^Y䫣qqc+W@=~nFd2#N7{Mƍ:77~<|zbP j9n?;3X y[|$`2BAd$HFH&+~zx͇T]sƉĵ㛕|{6lcm,8ptM8J!]e|b͝w̝T$ɒjHȤwu6PKA;vq\XPNyF $uCXoKXOiHNVphᮛ8(]Ҳ8JwI{U02[8ϕ =3"^%FtnD4_~#p9A-j0}^Ԇ~s?@*Gk<`0f"L:^ SZ1+5qft=Rd_/cwA| s?~nb9#a=\,7dRijVctY tA `Z; VPUJ=!RiF*4t/EUveX)%b bJ,ZA etLo c')<-hptu)9IJ6;86(9AB#\ [iApuiIJ&ݴ ھ{&+aO kwnԭ<[cvF=c=Z7|'ެ ,[S޿n磏~:YB,iYQdUo*q5t?1p9 o3 -.'hgHD3B׾GGd vpc͌zPMF詜kaN㌲Et$7$1,grO $F|lF0 NGZ}Gd$&R-׬Ҙ%-*D҉l-,!,EemBOr]^9`PJOse;9030uA;&l}@ibq4땀\[hJحX4ԜWKS  nf jLro(!* R詘w"&P#*{KeC-$ky4B2KS"QUPR[ŅNk-P,5$8B!Xvw$Bzl)!j3 dev63󵒞?د%#3K?zo2Ơ}U QZ'J;BtulSb}EAjGr؃ZU +ոފIť ҿo Owd,9*'zj+N=DLچoz%':]K 썍Ҳr;NޝCaVpx|K~߲$K?heCs86WIԙ3zt!dg1m]ҔnlYK~UFb[E3 ggt9ŞOs{ۮƈHtKS3#_`ލE:f0)P'qbBY&RyyESޙڧcuk6)HApV u;c[ S+tѾKDagX!/1dKQBr)IC3*;,lok!V*8PA0х"[ 4(̇ƽoL"m;Jx*j?Mi/{0T Rm{ _kBs"(15.^l<;/Uu{ o5s؇I vч%m_tc/2YhhwD+DU"$1DVR a4@BBTX"NAd *"_OI'DqAm+gT_ۃj?xkgK05I'4G>XvO|rÂ?W#hR҂T*BR2Q Iю (+RJC)LYe_~>j~\f, b_r=węOʱۜNۢ~`6^y4藟,ڄ;gW qV-r پB[, DOsWO&U,;ܒ i\yǃ%2Fz {|+lOJ}Bbxz{^x6O P/f9ETUCS߲!c2Ay44BH01ȃޛzq 64CٓsN7肓sOwd$JiQ 9G0y{{ 0u7{}RۻفG;_;hX_KBD~}s$;gIfϿEҥ )c*]aZR.'aLԼ(5@aU ! uQSdoaFʇъV#]_cO/ړPnGsL|FE(DԺ$+i RWЭJH[Hg`͸2*T%@՗C4ySsvV4\FDs(3CK@þfQ JaEh[(p [BCHw}s|Ԥ̲K!jiA*ʍ n+V)ڱL|DzX+\==0wK07w?0+hp}-߮^> @7KKDֿ}ǺHK?y+e4_x`Ǻ fp&ֳQ1qg8]1d]4X9__ E4Jh4]=v DtbFlx 1LEO4To\DdޟAb":hx*l艆jhLluxnBb":hyG"E#+/vhvK!!߸,SX ƫ+< xyۓʉ7j|_; )8n{j8%s[F-k;le#崊N`;ge(N4ڬu͉vo1Lj|]1@`ChO#GF򃲳#pEXtU7DsIq}xɤ2!0}F\EN2(cًU^)0LcjB,.pc[LZFis âc —) 780*(6Jݥ ˤ2βp FHU[eg!UԠOBH+/w|tLZrba_u_ e0?TϣWc!'&!U&ejDGIA5}̀<]pJ 쐖 Wl}?.XL+%gzH_!%anbqݞyڄuInc0}#UTYEѶbVqKR\Yu.3/ʠ\RCze*+eۭ"^,A}a^ D,+<%dU}SI$QL RTj$[KlV:{?M''v?jc{|ecVNݛ`g#$.iT9UE7h#q`$ iOy(TN*0RVѪ:8iQ6N 9 fN3qhkdRgr+E#{8 _2X{^u9ʔ<Үx(8鈇-yvP'xޙK=k1H^^j',<{g#墤mtY{ L)v>zfsB'ӌpc2gFcO̭3"Z~=d#ggW!ǔU6 aKQ DP0+N`6jC[#BAsDH76 hbJ=Ԑ`K7rAn)cu?RJW{tiΦ;ˬ8"g_b&Bs[\ YH eb7PMo&/"b谜~{s1lDTvwR(}{p/X+6p!a}^ le @mG)c/wK0v8Ř\mH2)5fAvTiFfNE0U-:]gշ ]^A.?xzY$O4.kd=[~hAFq$ oEFAz0PqnTE gdr8P^,!mttUZi3FU&3wGPQ1s|1e,WejdI-S*TV1(nI0ӒC`!Y 5Zr]rۍV(ྏH /@ђHn  F$ +rM ,7w#!yT2T\يpXC ӡGI@07&-A-)`-apE`cb5:D*u&N ״;5֎7aYX M? *M1r>:NL$T#C p90@I-o䍓 @%D5 gd 2Fm)JG-*Q^_dx8R)]TB}Qz"ޝʻ- PٸyQ;?S 1qaAeHֺXtA5 % CX*0J+PpR1Gԋʮ ӟm!$\7}y_{;6jnt9=>ج˻ÝT4n\*@//߼MMoo)&Z#n/13naD*yjlٗxw>&[r-Zw:]

;(X[:{/jgHQ2 3XWpǷV ,Zi {Q.XT tbI];Xl OLޕʻ3˻q$Q;;Ę6Wm{ &T~ōo\y:w?e-fOCTToωj@ GkNEC.<,agɪR#^{2(g;Ȇb'1 )ڛ.N,Z.)nqzf]_z_ㇿ/ǻ?831v|=/%ǣ/$0 Rurmx;5jy^R.@UzcHuUTv+iIqƦxªpTk!OK}_d^aEDe 5tʆwc0v.~=&l{ BxxnŃ:O{H7%7zUcA ;K=H%cAqY,|s(rw`u};#ώ<JPyq׻XG\Ɏ?#6a}v>l-$aJ٭P.gB+mJXTl"xI%lGdw*E g-emxllb,I|:n,1-|>71JM"6m%k[yBy["~@]T^~5뱽\_.%eY]*r3-^*C=2kJTLSցXԂsD^~溠n7VB,p擐cLJ7<{אBt3$m,jܵD$J O$kmX˞d[=%9b.JthJ&)_rpVu.g8C; ,, { 2񊩴`J\suh Sq}7Ͼ𹺜\,/kzJpHNu_\+vĆS dĎZm[l{?~rךAg68צRsMB^˶m}־H*+Qbp a$(d;7L<"SS%LT*T O`IWEZQ$D۔2."MbKzU6$q9 6#DxRP,(\b*'PێciyinGfIͯs~0[Uwo9qT/4L1E<s/uaʦ,xVbv3`W[u?.:BSPBOgsw h fuu`a8'˹Wilh$#-Z\ow!lj G熺 |]ukblԨbrDD6Yg-y0v2-vRsF!>8bz 'NsL )28T3A>.-| ]`f!"@ug\E(/*H?:7 $cDòB"Kq2 A5pٻ_E O/hDvR[ެV'׷FwTs*\/hw._}vz &оwc \ev~盏 /s7l?j2?gl%ُ/bmrVWWlqvz¿Mg_qm~ZD\-|FQȹZ7u -?wmNO4*4HBUuT7mtśGz#=ok;:'Wcs:QhŻDPY";n0Ⱥ7b@| se|p ,rN[~mn;.6SélyʪѤl@il5ff\v)y9e8b+^]=PGxP9bX*&61kDL=Wvy5hw7.Kݒ榜dF⭏Hf~\/ª,<1h|V̓'7 |f̿snH 񬷙]-j}y>p=ߓ)6oS񒬣v4:^N?F׷K\jH4"}: d[B梆Wmn.k[ӓlDlnL)%X`D>>fɖWW914But}&fS@+)>70 ~P5EYj}ղ./*>tڠج,-iO2FOӋ;Ar{nκ"w˫n2^Lml5y=+ߤr|9~=wo{*޺q OT q<RVݳ>{xF͞),ʦGxMj>inewyLbw^1:lx@ױ9hl4Cѱ5v~n&Qy{+]5̛\?rw")/Lo?rԿoƏ0}bw(oͤ.S9iCZes}dLڸ"A\)D$v#jIx<Q٣aϢr)_E EJ7Pt&AWuBa%ٜCia9Z*SKOm#3N=l;w;|O,Ytl SEBev+>nR;&@'֢oIǫdMF̱ȷ*lljPPn@R)41H+PB;ˏC֊Y#ƏHH50D= #K $ćα;BLew5hҰF TjZdzt''BJ& W;gTK= ]Z;SQ&)Oq7}b!tC ct^m/ZMl\Ř"^ޘF$hE{vC:eU/3;O7RVs4aa "̻ӹ}f(N2C %ֆdk{,&3vxP\`5wDO2{]J@Y Q*m4ɰNZT?L x9@56瘵&J"0hEfp&f*L^d>Ŧ\84*Qd!S- /#o$XJaǽz ;6ǹ={X q0ǹ7 }Ħ!`T㨨Fo| S yϦRv,ȎLC_Mϱ yHqhfZlzi%Wm~/PԦ^M*sQ[tYTLYsq&ISbO狳cx(f%Q3w Z &1іp~va6לj 7瓇|u3fp75f)٭^f󉟜-ٖN5'0V~P5`tUSɎ@nɐSw=]/8z3޽$@p4H$<6UБG^B䠒$+#{q*1%T<fP݁0xI][o[9+_pX} ,h,:vld:3#;dIysUdUd 'I"_CxU^ў8 M6,c0Lj DTArfvtQ-n)RLRg2|AƮh7"Р)N;)|6 ;7h<7SVZ;ɐ4*5rSdtNp4aJgRYNK! oMJA D" EA+`wUT|<{\,MwoU\ݹ 軻+KinҨԨ Fʩz+d5{ywWZ3k=xhyZx[]U׸XVb5}5Y^Rʭ9akd\:O,65uS电^5nJĹSk/=]rFN ?1 A/#QedHUCxV)Vzz 6Xp"n=H[QEY̪SZ_^>>qp=~+^:;=xssp[>'́8Q~O<@t(3>!y*9٤ZçЩ&ӓa;L%di))A Ø% \hƍf%ExBB4 C6) 8VY2[W0y*v}nؖصpif9z][ǚ(HxXx˰*d!2Jh~>>bͲraεzGx50F-*lt-񒼔^HܡGշVVFiX{9VQi !b:4LjbG mnW:-Qv! |uRsϬԤσE&un,.GRP@,r:pD]q,gQKy46pDwUڝRͧYގc-pӖ*Vw/%NHl۝pYXe=xM3^\})P͊LmNdͷi Tlw4sLLL&"4|#\j'xa(5ywܮic\}3d/ YnREdI%Y+Zy[ƤQ\s3-j5Ya<Қ"SZHo}' "L 0D&b]|օ*ӲIg $XSgt^ qZo[SpW;'PB::ic謿ΗV;_*ۀkjiM0G#N&g`PR9z "$ZIzϝ*a@?(R+kDOB 3VH$|2 d( ƹ D!FNV @ȔcB9@-_A6G :1㝊Vl)Z/f.$U :+drF 9-YA$kd9Zm HMuҫnL;##73Hҡ.EKQ!t Ca},;Bͣ@/'.@I b [?<s."ZTN{93Gev!A\, (!gehGd@)MOq2<,%bl#)X%MDYW2Pd$t\TU&`.cR0QےKf6Od]bfvL/tOG_oqbM*S-d,j>T6x s3.%T&;G 9Y4oJZ`>_E(1 'x(B3ϹZn䪣=>M)\ ",̗烄l 7Tpk k"5p&'v緷ZR"zM^-?I^^#,7 & PC =x/9@[󹝁vjr|lFN٦mwCIjfumM_S]0˸IZ${6D &DZ*RCfi,Ikm}|u0yņE҇˂XψP(F`gkrK -)qcb2)c_a\Ve\V a~uKH)Hf"PHͦdQgp9 1cr?[Ϗmr8a>m 4&]ΐ!ŬyHOhn۲Cڦ?^??t~8^`,T~z3iRo:#$iVM!iYܹsr9)Ru,dcp"fE>28oQ{Y:Qp"kG!sM`'Bn]0 8cSJn$pdgCoI܂,3Sɵ48 Z2A UpoH L}V3E8 _gGݝ/-:VO8ېۭ˃i.bO|;8$C֕jحw8}f F[ܞbuN7堝Zֻ7lJ$%lQnkIAz?ߟ;D]/?|Pha1tb8QPlHYn/.J~qp|˫rk9J\ϋ~Z}Zz[ϚJsv/,f){Se2X65iIe\eQ{K]Eޒ'BHx4'I>(2Nv" RƘ>%03qԤEN8P9e6rrHu)P&AII Ij"MMk, = D&g`|ſYaRHԚRǦW,mD@ʕПHNS<U2-+=? [3J?0׏h"H=ePy1 |&X9ßyWsv;{Ne5+W{НIMۡiǐY0 Pw7>VťlBoAf)hA1X|uOsMVxj$x//;՘i7^0 Qc߮Qphߡ ӈB֠T@Doᣓe,[ۦ]Yq(_ІO:XDqrxgGސ׋k%ޠr99 sa,0FjߋN2,`>kuYaZ/W45{-UOS +!*T*|RSrv,+,@k0²ON앖Q e0:I'Tv`)KxU!k:ېFQcz,[WmQv;gi|o0jn/޽\_qZdzN*0^AP٤HɯkSLB_:mulۆ\s1ʏyj6r25 T+j}'T]zx_'㧟~:◻:YkoA&|.NJP ?rU7>T-HM},oO mǽ~VnRhPl̃f+R`zr Nkzi7*#QƶC[pz=6 5W[G~A/UH QXsb ^ T&`)lp 6+{OAI_j >";MyZ4jN˩\klK"kyÞFKl5cLfu{:؃?mAe2h  {A;Ze\>eJ 5H Q :hraIB cC(btد]x0Fv9"~!3ϡ-^#V68KJ5:P%L?>3cBKFZ@֐L)Å>TFutX͎zwP2N9V K. ;+ : Ơp?1'M2Z;3z^{u? qAmٛ qÕ<.~u\\tF 8Xd1D2bJꃻQFsNQ͛,^ǯ8ٖp/:)CAUzy`]e>d ]}پzV-H,jk)iD]gy%|䞑}rGX~1yJ% rwTN݌3N#x}U`ۿ` ݶw[:PZ> +NϒPnnibZ9~+6}~~} ?Z hb7y(.mZEby2> ?{c +GMr1!9E)˔6T5Sg7x#"4m݆Uz1fw>E{)A? Pul@[3Fqj.k oBN |mwtQgzH_i^c;#Ń ؆>Sb1;x5XUdAG1+B LbHkR!@'VNr@Xo{)vXh>0 s6[0I;q&Vx՚3P#xL~ut>CEfd8C5ښQlvLsF-g~vD)ŇTB-kWQى%85 *Fua+6tN:'McV H?/!!{aTz235PLR hV[mJ,{:V0CkluMh%+/j91z`^:Gt&fhҴU2W X^<5# <W ι-z#ҵFL\gf]W^fJ3s@~M;\O%&%[9 'h FSZ^A[K|MZ)p$wfdL'wV5{}_<4o]"&yw*~Ǫ&(fvlSfI6VGa!Ll:<smЄ (q{,buώGfaS%٠*9?O(ZX*LlQKk|~uRͣ7SgaJ@| S̓!ULd O PYSXQS(`DL.ħv]FZk5ƎϩG8cS{^R.7_ ѿ|IaNhQ+q7olp QggR";\_r7paĊ]5t:^r ft=u^lXro6hxp[;p[0[k޶s)iҶnۅsi: j[zJ-n46+לFۢy f8ArS@|c;Dek@fE@NBh,YIma |xƶ(߮f׮`;Ԥxfv7.3qg7kU%ν(p}nYRj€9vbz=,{-a.i,DOdFtU3͟.hw{SڀW+O>HCu}}"Kwy-/-O?,eGr8N.\>W$/(.(_nKXh΍?l{G?'TӪ7b1KAK Ǵ҅҈#+Q5Bh idFG[kq$K0v4; u2>z =z@+>}4W,>&1HC5l"{j\}[}sj[>9HcpGwm{qШ;/bMZ2& ?$X-~8`m'U?>mo276_S(>8|g &_}y㌸ @{Ȅd?cZTq)ѲDo8&xf-v'}!bOv=Q Tɸ5K+c$RϾm"}Ꭱ:|~ZwVz{z6$cCF,OGJ>tA;UA]di<39.#;I0~L19MǗz*%4&i޼3'ȍI5JxI1%ʊ"ub^kg.fn@dG($6iuSpirdMFgZ$V٨aN+abһ@._|*cԯ,@+i3HQ:ezʭN(EaS)(Xdꥶ6aj mވM'kZ~Tapδ RU|MfFpoc DKnAK&/&]}EZ<wu;̧_KKυN Wdq 逾L`߿8HvΘ%؆Mxx~['5S?Z%;U骣?1umvu\r9y\Rڑ{*NlƄ;GIѩmJ\ ԳgQaJ036٣G`Eg6M%gd6a/3c+ \Jpz^cU8^iB&Ceis̸<4X%Y\2#A;IȐ6NR818fj൵['lII?SNxivY&Ksh9 R)#\i#649% b;㶹߬炙 oyqvc˜XpVY]xh/>htV*A+DmTPGcJnOd6Ң), a0T2 R31쪮NA;W_]lJ;| -"4Lo6-K:z]S9ȸTI`CPk5B%/FH9uY\7TFv`@8 Q&ƩP}e1<T3@ ]hH. 1MEw&3W,DwZbPLL-FLB*ovbp6V '&V cy$Z5qv[ƈׇ)6Ѷ.rjkh[FQc'˻(ҥUPmM\QZǶ g c~}}7+HA+V(]1[B&fST FViOw>9gj0$c67hʞt\ xC^m4鵊mM-4XWc::LGdhFKusn ({qdON,`r޻J׶6YD\D%Z!ȑ $ide+t1@ۅ6yՅs=r: >]!e}ݐE<5?wS};Βqo;w8^ $,GZxD"h.*zI&=1CG P&\cLp>a+"(I߭1[<ӕ k5`Ç\H:p'BI|!b]!bõ=Id ғF97z.wG,]uҢ; ֭7w,^r @AJke7o6ο: - Rf݇`[IHK*LKEAe`Q]}MJhp֌F(}C.WSW|=<G0]xx䪅G)A8J`P $g~b~h,V O?Mqy8+= -nnQyJl$ˤq,1kkI^?&Hv/D!pd(dy!:msHYB]7y#.&&s ` cNhS-@,ުCu{x `ggvMnMJ'W9K:@6B4>s\Q{P-~jzcÌhHy9OQyj 64<_3 J L!\d1L wC8@b*ՔVo<3h_/ BˈɲZI$v2fwܥݹ_]:WBr_?<6Ćyp͟8̆ua5iq~̿?fڔnnuȧHЇ^ ኸNAƆ%:%w@M偹 U$ |a)a҅y-% 5QY^ ҂tt `K.w=1 L AJB@0Rbu`5x~E1:XJ[-΄T$w<%xNE[\#Հ<Wɮl# ";PT+w@b˃MnN ]p(>P%lOJ&`fYxOw7i Y|>zY=<^Tu:"}ZI(KiefA̸ l[;(=.1\|})a8\֠%k増=Mhr?VM1}ͤezo&' ~;+rwbq2eocL-z4A@]48FZ*]BV9MIV]>>w=zzɯUwnX. H4QЫ6 G*1FiWyA/25gĝ3$P5"W%Kq \q:N%F-)՛>̵;RYb=?kO0p'z+hFqopwC[{ܦ&IEB— Ft.q.*,D?kH5}hPba]%UdP*YC%}GšJ8o@%,r-7x7-OQC Ă6^B=u5IF+Xˣ „O#j[3pfӰ朖 xa@4t@t_cMs6ZF1{ОbҚ{`)Eԭ8j !X{;iX+s8$TBz[SJN=W 7H߆nl*YPr%u<-&&;ϫJ1WjxiϪH6gqKI{&zr M,R8cj[^:B> np'xz[J!Y)$QFK*rd.͢.4$o/HVzw!bPLO8 jTJ'UiYhZf.Xdm k2̛4)a:0IV0]ZwfJ3J9m++/:Ĭt{^+*[84Z0D EӞ!%O/qD0238bobC}D(' `~p1p]/'W!w_ԋP2W` ,zM c4u'rL wݟjEvԮ!Gu{$ࠍPH`f=d!nz MM=45')w4uwf4 &yx"5|Tuo0'[<2SJ~lYH_J?͜yU8$co零>uyzzG7=qfsw=zk㮴Lݕ[Jjnt`F ;vʀ̇^*ɬTWVFֲlTކMيxcRWWׄ{/i7excvE*oOP2'Qj8u n&[QRUTGUJ*(=2&D 82UN~QGQȔ pr ) m3A=%ě0ZB?˧ޟGZ|̞?|u 3."}?/?[!0L7QOoNK^ @KR@B^ GǸG$QX{-H,v.OPڽߖ\ mIxBv޺K;ZT]{L}*Uқ00ԸBoffoٽ_X)u2nt|3>pԌW]Lډ.{iĻiPQnSU9M/էTgK8^7^JZmVzi }3uS!鑊L)ӤEo=)FWViXi/y2d6{I+P<6ZX@ dL PBZ] ƸMmcup'|~`}BL( 6RV?r." Q%o_9|)cUdLQN :ݶFV[$v$o%ә=퇱&L 60 I3^pz@SBĦlSlNcqS|՛"qi@ݟtjZ[J_Ry(0IXsTRG+}V߳"?%?הf$7|]JIF~S! JR~}Hm:]n<(GL/9r撉ٙCA&>綕7EGTJϺ/e%ڞlLH?\z?ΩzR7&UPj@Qyj;'_f"u{YQGHhʑzt{LL#Τl`1ȇBӛ,m+dQUi03r2MHdP0}29m0JMl79$H3H{gW[u )|$,Q(T1c] y4ǀ*)EzT{Ϸ60Gjj$sfw9r2%:IvsfD H rk5Ā0 ,͌h6ߎ=L j`Z7` 7%Yk4ۆ,3bJ.1ُXg67rҗ$ @U!*ueSyxHv/~ɕ_۫4iF?Jp!3}ջ凜 )ORL rS!y.=U9)c${)xnJzx_)T'7ɬK=E7DSD**uRO1a)g ϒҲynlҮ^ }_r)u}{ϨhP*Jpi #$- Grj7mh!EvaBZeĺC<vSШuP)X<\ SVf$[dVM6r*5ϺvBa ڽ*ƙes9ԇ+lP&b1(Ƥ?&ucWힺ}tWd꛺NP7].\ uY&=Oi~Kx AJdĕ{ ,XKZ9-5Z-Zasox).ΦzV7JC zv!ma H5I09 ]rA.#e81 o8L{ T&W3tj^E EVacDR111&;Zf//i@Ct\<@:c:\JDӛl%Mö)k1|#*=u'dqLJcL x[| 흷DI¶<.,e(]s'm6 !J#۔D((qAJAyUrQ,E ZTDrVv0w.b^j_Ni/AUp,E`a)c&0Uo$FU ui*mL&%T +8"W8/ZqSz:QHy֖Gc9S aVUe=&=z$ /1c&m1esbXOK%5iɴ0})3$I7ҦxY3io(jsǥޕc6eMIE8'-ev~zʡݥ/=6uO.-F xpu8}&noVۆbx J:N3ѡks-}"%Of b8 }xnw\nZ^tPޯA:J`_YXzupϐ:3Sj|*މzGni`ݚi{Iڑ+^BF cfe'T(SQCj*ҙLiw<aT v^; GJ{%=Mzd}pW14.bD徇vD.p\(FQ@B"f ኛ B Ծ*lm>%o-ը cߦ_@<(ǀ/JV7}Yf8Y^5d\3[;B$T$FO&E(V0@("9޺,v#%az54VT1gDp,ϑ~Ǝm79%u'p)h9o*BkYT?ל2ЊT^R*)"/V-s,%bU$zeЪV*Zg^\U]*o9!My?HS:ܑ4h#9Mm-ɬOX*T~CxMr4Xΰ=3RR ZM_1b 8q6*PL.,I ;˪+vE1= ;jضAv lۉQ]4J(!;ًwէ~D!M{;;U H?XW=DLS>&:_#!(X=lZ#(#[^sXIŨ $'N*g0!悓ˆH >7VMQ7EgR{4 k3{-a1iGruUuD2[OL]V׻1wŧ\OctoC3q XTʮ} -ogPcyLLIl@LcqpCMn=cRragڱ @]FDVM#֏lQ)`G02+:DIe2_!/:.5_.euVJX}3B;MUVY2а #)2 yIL@WHdǾL]@lDٛ&ɯrBtu([y0+ԭFM@ݫ׹g|«08j yOF! }XwOHcIt% OQ| C&<#~%B* ""a-Cć(4maA>7.5Nj)0ǿV~q-ŝ~9~Zj0Q73.#iިqܳq`MJ!aIϹ`Ա?ҮCa K d@0ы85AӵcC3܃C=A8PxXmfPX,1Bh@ )IVr?xu2LXr=̆ksT3CL3p&EwD xi;["|C]dv=8gοCx(>'(%Ar=] hЄ![= ;78 ΢D „e`H !v?yQ1dD V8rT0ƘU4?2쌀 ϜV=6џk7Fj5^MYQ/u]ULƘ1c7m޼OfPESK(9FM]DQJ)Z`$VU 59EՔ$2/( 8ayw7iյhCߍLȴ{^V*^Moou룫϶pɵ=n;·۷٬{{ڷ}^_0+g~ 3Y+g2_,7}xw}cvF {tawp)B_Odq{}t $gPqqy5mT lHAzdҍzT.LYv=ax Qx=-Jm UT{f)>/oi4ʪ\kdv}[ P8[{;j6{^Η4lc6m%J2iuWI"# ښrOW湿GwV*ƈ!v/ק=[FeU?:.q)nq߮q*+-x|zgnWo̓jJ]vEd iIH@׹|yݗ|nR|;7|{lw=4eJ ,nǥޕ7gbi&L$aǤޕc ^dgvki"RR%kP 8/Qjds4 #Gd䅛]LouBrsOF9AkauwsUuQU[=kmqKBΣt72 ˱ su1K6qzv]<pN%oNFqeIXI){1ɓ]鲻1>pV=zH DJ^?/Gt8e;=qxZnk!T|%i{ttpч'H6\®}Iه&)GsߊqqJq "HeX6Ŷ JEkg޹|mb0'aRFq(cOz$K*v$ r~':DŽ.$THe蠛J6ۋͦ2"k iQ>_&e}?"1B"a5 #+NY)`P=A&,(' Ly[A5)B1[ .TPժnNZjK9{v 39bU:2s6j*Mgu dU7{ |Gep*- ղpV?~xܱqldbuqw>; M詌gM^'[6 n:V4{:]`U%}aFts_߯;  n>|Zb1yjAV&{}vK/DŽnU6';EB͝v$׋>pR~0BKhc]ǟW>èG¸ٳWL.X1P9rXkpXgI~:FgzǕ_18Cfݞ> z0=sӢAITg'k;s0}Ȏ/EJZiT>X7V=6WEb8avw:BqRZ76DbIzJ_lÔ>UصɸIbF6xwɊ/ צxvk r)q _lR|nsiZ+eQ"ʑN%04ptJ2Z+TkAT.凼|k#DutAd~ wsaIsǹ|˭/6l̺ MzxDwx5Y!Tbj4qHR[-84a6/&y{mPVZDΨ堙4-3!]( ߲lYl-~~3d"f؈ [ "2pDǗUDŽBcj /aT;|.2} @͘['գ{=-dnIm4^ a4Z(+,ɥ05Ī́ɤ"Y©:Y#mdjT]S:c)S,甈 ˈ2S FQeAd(JQd`y&Ԃdr?1ibf}hy~>s7{\u_==GASK|Ӑ0; S~Ǯ껧ΘR6/G0s~vշg@@&%Um/FlTLf nn&/>W˧G!ɘR @g;_|yg<3fL3")J)A%g~Nq,?-y`N&9gƣcm 9nAػuP-њdC3\0f#G{Ѳ}WT JвMl"{EG+eqc3w>s]Sy)~,7wEdX^mNq`A=$4% gy8*ZR \Q2k'5HP4SUxre,=҇%>8;ErGw~FG+esk WYˆ w'L(l\Y ]\_?$=_!#}5g`N .skD[nIV6s9:*VbOVO#sEuPLJ]d8%$pI0[VL-稻fT0+3$Ň]&,@1J,Yh?1PW!G2PIaQF3eVdE [JSMthJ!7xy41=hSHMl2ܨd@py[132K_E 2S1N)G&ՊWL!R&#e/܌ENu6ޞOZ,Е=qcK:d ;8-rʹpt&")"PRr3@¦V(#hA4Wq|-9H.@)=tI=c@xN'szJ+Q!| /YY4__ $qdRIԖ$7()xQaAun6h" od+8] A^Zn/&0y_*J רJnpZ []Gvwh,P1atT|[(>rt?H#yBZpx@d;l%򺭑Eau0: #ƔxQpCPB'AT SH' $|M%p58߮*qq֝!3yhf֐FVv^_?FzkL|ӫNaq<]η3W9w/ ZFz\Iu;qkrӪƳk09_ǩ 4Q.9x]㕳/\|`@PNuu,Qtp2Ht!$-'EvF;ㄚ(10^|cȎ|X@> $9>djjmppi )-Y#楓2׉3u/RLl_\Zveԍ5#Szv~6$w%r) 89v9걀{>Q3f150:0"Qww]V'PC3.y~cN_+aة=}>CT 'yM!2]z.:L`TL[8/^#˧;ɈqH*4?٦~ġcrZN so>$ ehِƋ#--٘5j3P#S2MRZ͖9C 5_?jN$gxX(RSkqm T!,r1Tw aG/+J>R~J(4!"EBiB3Xns#yr(# RO,HM1JYk &`.6B˔HmD*Qӄ)ۼ e6ujHU`(>6F.o; Ę(Q6AqGQz, oҺϒQaUfȠp1ra7lrEk,E2%k]u%_ls1 l9^=}jb?ރ$L#Tǟ+}r=Z)BUQ[4H;xcKą;Q{6^i<&w=Xr--ɢ ٶO!̶ѱO=3<Uq_S7N>~yF]2t?>DK KF0C< X!+c\eZ&N^7~bq ڨi*AVL_Iq#A-]i|{Ti.w|m?OvC׍aƵGOs,岟8Lh8L\!bQyqs=*pa<ݓYYOMy4=6N 3}DQgd'˻FEcoKAVTuJsk졄Pu.큩{@-0G+/8ۗl}3ڜ/{-_YZjy.k/XϷw\~Cl>|ޖiQPG]ttF(TmBS1!ZUa~fPU2'S{ !ȂS+MbOH*N-^.5ES8P%䣑]E `iW'[Y*mYZ1dM'Q`xϦwe*Os[Z7CZwEdbEG#"4JI IaS|aP R$ ig 2()xWr;B(l^p2)Ü9 >u6t~smA[&چv}~ӫARz^s FA6ȦCnry[˿=$itKۆs*[v8}kz8qO!%D=9Tx<8z-WPjO8 tĐR!Nk5X2] F|PN 0bFElq=xN$gy?L](NX|tzu.nf/J4:y))c™I)~4zSnKOE@QATZ0_-ՙ—%R9c9'LA8J*XOV5c9aR܉@sڙRZNS맓BoddNUjҔrEӲ[poK2=&DAϨ/N5ߖ i,ּzzT!*qY~4%}yy^j> {+'%5/0s~g[*=x8|駇HT*\*F4Jӳэ/_s;䌃)^P pS8ݭaIm D _gi ]d]w۠@Ҩ{l9!D5.C]{zKCz5e>{DGwL񨇓f\wb[Xw?=]l8xԆq#7Wh'cNXm^'?Uh| g\ԄFb|rM :"=kWK{PcI*; M}X^I+n]5cEç)Djq3I'Ezw $JUk&,+AY(a1OM"~Sھ;W6 V j?Mbe7Pr۠Я;E5p:7ygG}".Q(mn_VuIanpL3R *8wwԬUmwKV|Ae?Vk`ߠ]oPA3T$( & + SZl|[`z;P#nU}\@)w|;rBLw (_ H9ţܗ# D A%wlFUuآ`->@Rٗʃ $vH ~| ELaDC6p%8cQ|]vPPgd}1'J2ӧA;pQi]נwsE&\~.kyYndTnsR?fo@LO _^&w 0c8J̦5@dYp4֙,G B)TrA-Tit4jOxy8h&-P| ?{WƮ/EJ&93|1O=)p[\qlזsoPd{lrZ)qYs3Ç;O_BKV _}L6sa1^fmo6^¤Z@ 6&)pxlZ-A\Dch4O1Llm5mtIF%]i*B0ar9FKo!DɠO\#ӄXU05"c?fsbK1Rh@l< GKY»uP2b 3e9' ]]\/g6q ng"PB1S껻\ˤB_ V2P܌$BN_ťV#`r ٗٹ]T K$*Oҷzr|Rh0!K~o7 5¬34qk13aQx쾆fgov1Q ]GtBw݌d9 WH:AJ>`>ϓ}'O틪R?/lH&xlٙYYɉJ7+!%{+86 H0b ia!N;!Uβqkri8dSm>W@2HS; "ZYvR+޿IEq1+ļzS&e<"# R;kU2A<H6GX»mФ?;dA> }Ǹys_A]lzat7F R61j!oT0vWy[pcTt Af]=xskstWb [Q N4RBTfDoa}RDobʼnS!gޢZE \ Qَ>e^O$W R鎩بv_?%XA1/ʉɫe)E7׷7_N%!rAz?ki53ffQl"~pަ^O9#3/eln**M8,=1&f$)-Jm1%U2ZF R`鹸smH[hj2W1̈[/]4O\uƺp5og'DJ.Y2}L ߆c.*e9;=BY%D~ r/΢I˓9|bkbʯwiCdzv\+@ƓywPrz?5-v?<p O[G^ G\0jbZH4ZHpve|MM8&[ardkWѪmz%LǷv D\{IsTXR8:J7|T&RBLW{ ᚏ>i}-8icRp;3:4}-pxjZf;j ivAl'j>z  :GM Q!| !ziGEj&Ǚ92B8Y}U_= xn]L%1nFэ7<%Dt,EZCk_ak*O-\ ~ ˴J> 5}M$짾11cV ;ʡExtaUCH-)VW0RndK <9]ŘsJ,ZF}*oI΀97C.,4p卙=a70 ["{6"|mZWK~$KQ;V%'X\θoA-qB/nwVSJbݐܓnUx v6w?5DR daO\7OP079C+7Wj6 ;!qC=y2d4ϻA 4j6{[%8=@ b`vıPi+_TG ^l%+O#*Ph`pdr lhhiꆇ))&\eؘAo{X`M3L>gY87FsVryoI|kPSiZmh1&8Fg [n`̬q5 rjJ fpj9HVFZX''#@3m&:۸P!Q ӹktnǁHs]=kx.rKT^E%QMO#D0M_O%gJG쟍GlF "&T[#mѿ=m6U{5yC9S\tedq9nRS(m Qyﵲ5h,WnH5F5/Vٗ:s,Q"|{p=V_DX><Z<e\4(C޺X5?bj%ca*`g!u/馞DlhƸA'h.ޔj=S2b{9.$bx ІSpn狻 gN^̮NO#D_.3QTUt̘=AA۹ ytu Ɏ,poPQ G }2hKy j}(]'^ҩ~@[qwJzƾSg&\_8;R\v: LH0BӓtfF=I½PbzJsWyjc0D=-t\.Payj#,5D`ՕR !8tb@^&G߻s@T3\je #[_:%3ŽSǨ3G~>='wQ:rɚ)~p{V)bqǓ^ ‡;/ "2>u 2*!m"x8KhG1[͒tJuڎe`:JDޱw9}X/pe=crUAB~qc Sy7 j11wxÅQ =3hM:pwS bc:(nBab9ym MtӦ LLxgJfJ.94ߙƔ\ MCW/\))y\& ⏖Ձ] 6#+^Ju݄t ([L˴H+B;U 2?v7^.7?FG~yCktPﻛHjWnjXRn(D-J=*΅5fqmT|X*B7~mN%NSYʮ,5\7(xنt]ׁ l` X!'9D8 zC +DZn]4PϏ[aG6=pnhv.ۯK~*Q˜^bf_Sxë$-]tkcC(f[i;Zk">)DULu)9w.'ç _O\F0Nkt@͂$9l/)ݧv!9YW)Ʒ'TQgfˡƟZ)HQ~k~{vWRH]3%ьɣcUW xxe%j{ʁY9=ל/d~_K4tvdz!EUt͵/Ϸ?= |q%1dȝnrgF1"6e,2&0m.(sDiф™)hS /{tzIs'S\d{@ta?Qy"Rs_\)c@WB(\D(L3eBh(ף+P. ,RVY.h"(yLC+ yIw d(%#$zy0#q%R-kӬ'b.Ԗ$YkWOOb_[QR#*yƧ^xƋGBq/JxKAF'a4 +8D#a3Y\P"\| M\VF'AIT. :~]R-Wda ;kAtQ@92p^p[ !os)dujģ )bvhdvFцmF3A a'YE1z8-_&_.f|psy#j=>m0g\loaK@J}/q>+/wwe5&ka]pi<$rF$+^R%seG 9A~q<13Qj`o>88X7&0&/h z{φu1/Zg)o\Qrة~erž K]!4j(bǤe&\qOZOH1w^u0k󾂕3i L94^/bRNxڤ{}#:Pp+'wi! (^~үFɏ,6~`ⵘTkތYGxAJg罠SHyֵŌsY{Ř^*6ͺw<]#rrq~*X Y\6}WUx)ہ=6c &W5*K*߇+2&[TaDs7~yc;ws=0mJ[·w LVD928kpbD %.&XŎrHMJ 9?Zhe8y[H 4cD:,}fwr_?5#X=o&c[P8XPj|?){9̘:p6{n>_u,~vveqwsx~~qq=>Fy6/{qEq/*B?WPSHr&%+abQMҡ݋ދI8oDa3waGDi2+N cj D!- V(KAdJx RMx5lK bEc~;,1?7\ӓ稾?Zm+^@[ـK&,JM?}0J|wa/hDڊX$ZOLqGT6L2,CYxO 8D>q˛̝s<|y}1õMyow!3[f5p?„3vAlInp#`vGqDS"&UX=1q! y9tJI,+.KͧgjlnGI$uZ-Y| mpD#F:<ǖR.xٰ"f[0R>/}]^QՄ5 tu˳ $+LX ?M|u/@t-/{DyU޹JVK0!^+%쮖}\hybJNO!S~@0Ŋ+v| ޜ8[91jMdFG,&)J-$:7eB$&Hx;.ÈQtI6a֨!~ņL`udB`dHeTsJ61EJ&&JXc#c qQja#9H,$G 8F[psgT+6cj)?]vHМsisxI H1CXaf4M$ &Q[G$!EE}oBHiqZ;(D6PwbgEcusWoKT7j~SOoRL&393u"XS*+N5G(Dyi)uD@T_O5hzlL|{2LP} Q>IE7^eeڲ @oS"lsgT ۼư t0`JUvFX"\3jsx -ፅ3Щ"Ǭ>yk+eD2z6Jգ0Jרm7FSwk5d|V^+Ⱦnm=S^ٻ˹ ,nqʆ PJpUIpɃSoÂ̛ࠇO"aZ0/AD# g8@"G Cw~uf1J3 wd83`ʢSfʢA%Rv)K4eIj@].j]%z%s_.no-@L$8i]e!rD^4N\ؕ7^Ǭq'yT`׽:Q,t^UJ犗jJ{}V)I^ZJ8PTRҤF)5y_S<6UJXT#kp*"Z;U>|:^< K޵ $Rj] FR Ψ6ڦ*u1֠T$DQل",o09ᵦs|W6#)pbK-߂('ӿ]1 Lvc+UAeۂHsJֿm oA[ iIxAfWa 瀗fbQcË w 4ٻ\$'O`3Ņ -)9 52rDC GS7?xJQU3"3Y \|d. Z:h?6weƙdkW.i[C%W6W aLG 1]Q4"ր ⻶/y21uhH_,$V!E=A6Hխ3y,7T(2z[H+L7~*+*0 nqJ[IUxm&XHb`†nob#CŰ&U6׃#I[ 넓(\1qzVvAfMk<O]T֜;~w^E[sa>@L1'6#]D }BVAu iT!z8\"TS.buḧ!M12e\KX2L"Bň}BeҊƖ*5XT KVJcjlBT…ae*9ف죿;>.߭]go 7M6OzF0F&_~`c}qC}E6e6Z<XGzvbj4N>O4 oƷ?^ĹT_~zg$L)FiH7݈dcyƎ'D!l85"`~Yڇcn{lPخP|lN/0 w dD`"ZdUk΁GbKZc{&ìm" 5,j)x4p;ER!ۡ稾?Z2iRذI߭]^SKQ}H@vZTLohAS:'E4&r?WX^1/ZyЈj6"sND< p9t)x?{ܸ/:qΜ5gw\$#[$O2? R(ZA$[SٙL}Cm'sFY¬I2a0Y0BI⪤B]O? aFge:}jn+_4%_T܄cנ.A }]S:yPPN\X }wx> u 9@SN?5)ў!S'H'Rpgy<@8oƸ&zӫh%OW)9?{<;WV/?$sFyA=t4?|L9&?Hi^yӼyMjddXrb1""Za*sye2Њv$H9RG .2ڍ؁.Fbb䣢[ރFCfY9Z^ʠŁFښQZX&R+[ bҁH-X1ZcLLۨGeQk WBL7H,;fʉ2ˍ5s5WTF[6sͩL! }6Y~;GiIzҌ1#y|eb}.;w\̧ꩋџB$>8gYg52e rLAr1 3,i1Υ4Is%gYg)x[YB'VbTr_N#NT:HP5¸Rg#&b3 aFJaK .8(xaia4FJN[unW1(!`nmWsO?DPxn^R\)ԠƘ1~Eb.$(KD9Z^i@h1sl~0R[_BڞK2X'$ ZQ|ƒ7Jl/Bgt6*}w{  pMbcfS$ko]:]^E²VYn8QL=<-]ԓp;54(N5}. ;uYnp- ~Fj&0%Ytn,.:o>h٧pH%V8ǂ I"j !3!{;w%ӚVǛL: P54jY"ZF=i.?:ݚjм{?Xm@/  1ZSC7~kRz5i Lcu稾Suo |eXf1ۦh^1_KV,(պ#_U.j3gt7p͎AZV.4xznMyuF>v[XL>=E%Z]%7hYA/W C]P)1L\>eRfjyD)rvJVəתV(g[VJ.FDH5Jv,&ezQ?JK!W߅c"'LI,-J '] BMv. >7dtmUwoaGPT4HE;QG/ioZ`ۏS!78ɇkvh%DߢMpB[NUz{.Q6__ٽB=Ɓڐ*9v`9 !^7ځ ֝܏|EaEENsUZ1g9RVL8jI&"C9N1U9UG4=ݏءM`^G*B I\ BHkmS}Yf:8 $(2ŸӺ*abL1 `a",3K\.ry?Y, 6X[&ƘgnąR.O6E]Ryl\(pW[Oew?w0_ww.?}1ZG@­3GL\q)NnGx`h%NrgKVPsfds"! "9J$Wm4@">ˊ,1*75<(B<'@ݽ_M?һ,I;ggO*F& #?83wnEJ>'gv;j62|YbƝz6;[􀰂ݜ9&PRgebұGه vjh4~>Pa2%SrA`e%e1ѻW(ۏΒ/eʇoһs: IE(`+de #QEݏv%2> j8- 0^Ԧ_ kR?M߾jaD(D jBx{!$"ͱL6d|i!Hn*A*qOdjXXߓGPmB >:Hۧ辳đR]ËOJ=s  Cʧէn=u58XW42c1ļa3X'X V)rnf}T4+(+<0@bRU<tb$r>>L+?TS:W/0,0\hqs0sdiNq9I!ֈ1rb2M/2sC Ũ`c1j׼)*Bva=xksHpt p&8,m40MbȺB)eK742%Qra1ʑ0a8cSf)E|I{cr^lc^l9w'ִ# l QVR$ZbqG aNY &4ZRXi.mڒf0(ٱkJq:bܩOdFDG{*p"&gh$Yn]+k~ Ɯ}^VwX9sw~,〳ڒٍ<8Q>=UvFi hJadž+=ђ5J/v|9%3H0@!,;O b*J;Πyoa{ 6C_ 0έDCo{3hAR{kB3HZcoi=Ztڶ)U]/|#rBGZt3!Vz;É.9IhtTk̩륤pb%=+1m鲣*:gcɭO, cƐlb50rF1졊 9IEᱵC33eI%遭׻gʺfv?-F?E৳:#"v{_ngnQ|[r:TY{ymX<Mp@,~U}rˊ*ˋY7˜FĻQQG+ӉwPT[|@օq)IaM [ BL'v&m@~xp-PօqmmS^ :ax ZȮo.pbp ,aJ;Ogءrд{7``]rEQ}Yܞ."X G'8c E6RHbiS !L8G"sJ"j3 Z$Yf0d L d83Vr+-ՙ,7yTC"GPR$0-ve#ŰR!AXsT_֩RdGiO+P XJM3JM.QD$SiٹH22uḧ0VaCD@ =YXCA95 d &z_qw7uw'&`!A| -,@1%K7rY SyWB3Cv>#~d2recp67n )Kyc)Ds\]^@61iKaG%A{,5]n`@X4@ZNFGehT?{W۸/ rUC0ݗ 2)0xȚ#M6EhI0&_uuU8Ғ=%}{ E}/_Yi޳Cwzx10(t<؝X>?;rϟ'Heǚ\68%gI p;5h(%޺f(qNokSN.%*4XSyU.\t$#s P,^em޿sjٱ2x([Z#:*c' erxXZ'-{ܨSSXHU M~ahdw!_N⇇ۛ؉X2mB7ѬO}F&~a7/Aѕ-ҰlPʈ 4@ bbp ȩs *Uj'0|0sݨ>#SI, kcVܥa( DT+uk]%?s@v {TP9,z ݃`Nmx\lUvq.!@GV*p.I9yKyX/|WЊYչ!WI3c#|p{ğ;] ;NRㄤ$qJn-=_̊]mșg[}yNV7 dz |ʇj0bP}=]@?%op;,uR##FNzBQ=څ:/M*q78@55ߴ -oL$JuV`d ޺~jpksO0Dck,уh]Y]=+Eo:h@i:z:{d7*M?G3]aunSWs.0>{*7x<9ɶWc8ҟ,J'wfjVU:u'8Q̰ʭsvm8%|uruarI#Ӈ,͟3Je^Xr=SkNկ~CEHUF#=x\05#ŕ̌e%=cLQ%χ7LV"0Mcc=@m)=ed~l ƺ!o[X*-[o)sWOxyH IHZ9(Vƒ[M! Y\d(P͈4y =}-s}z{9)$ѭdWǒEnWJÏ??~䭘DLؕz؅,$Cwnș +'݅5;_>m W8m|Ilj]dVvΥZ>>EAZhbrϷ]mH /Sfva+ײNôRJsU;F z&N%jØĸq\h{j[T_թvZzZjKKu\Ah>TqykT~Z*TMc*= ORi8 F-=G-E᧥(fðS4ΰ>k-o׹=7:ޠsv[,.J*"."7 ղM`:V$J=)ҍO>Kt "Z?e I?QvIi1*&pESi2\nz;@YgxOBHIUG$DŜd$cK/oi5JH14~^;vy󺪟4;ఠX:t F0TiAR?)n@Dž}XƚN~((κ7QdIJq3k#<(cZdyL'24OQel!Y30^sWOub}hep#DAad^KnPhvFB_ŎX1:sh-oOHX\8?OWV0PG3goC䓲*4ef~"Ԁ``WSn$=߻"Xy*^@Vzk`?Wu-l%8ZZVœn.~rT3'٬B9 1LGEgipݲ^\~Yr[>o\Ƽ 1e֛#Э&0IL`㗫%k,X),1i=MZ@vS! ˊja9G"=7RZZuJ-GUj\ZzZ 䧥@e$! <7mpeJW.9n|qgxyeG7v6.pl/rؐH-"ig4-;y K1;QaaXHt'O{I{:|NJ IR򋏨5٩pnsJJt+~(]JJ2,I@6MvOHy綻wz߈l8Gs.dv1 X":r0FٹUN~S;IhUȴ۾tZC a_*B<Ѥ^GuK~U'\ڍu pTaKk'Gk,:6`Oy Mrtu]7~ lj߻_~[{x}u /\ xæ+ DIQjVYzWD"gї|:C"'@Ojn֬v֣5by6ŷ bd3ӎPX|Y] |)SWguT®d,1E.Bh9SI#>\,Q *0€ Ѻm1t-McS|-n~u};hc_#D3iSw6;}5 푟.fѤ9BS njfC0lHxsiޗ~r6|Xz7tlX1VԠqG=÷۞`U.ޮs8Ѧ3')J߸Pb=ڈt&%mD5e:@*C]s R yՂNƄwŎ,pi?Wuj kR窛M|BZ4{X) nK4`N PݕF6YzL vۼ<}qyAC\i[UEQoLlm2[FiWbPR`OLGӢ#6Ga)=sdO7v?7ݓNW$3ij(DCc9IFgbB)Sf:ODQFR؂𔣘*&W"4IƉDV5/RҌkHUyD Ԡ,$/$9 WvTHjva?|WCnMozi+MWkSE+Q/e!w^ONA~%+D]xEFxLؤDQMVR)B݌KO-f*C>CJԽf1x(Qo*m< fĹukBĩ$ppgVxCɸAh+Nډ6]zzN3tþ+2Z)\l( SrFv!Ja4^0w!,s14ý->347$;n8'Kǫ@(%y(h@01Kndb߻"Xy_WzPW)߻yg[*-lVǢyC2X!DHio/.ċ,\ oKxBL`uzQj갡OŘ)5̞f*Q2N(8g 7 L)X\(*~TLVaxWǮm58og˲<9=D\_uGUX#L7ti(o~Tk/5HxZ[r8ā:[4{˨d.ӔLIGp\ya{֝8>ޚX,,6aI2s%KkZ_D\0~r#nO}/v|5%$ ajh фASihbm,18v{q=Z벴.5D 4H'">( 6&1Iդ쮚3֧Go%I02B0Gя;x!L)`Iܓ;8<ٿﰀ/Og|1)mh A9TnO3\O? W5pb\] ~B "+8[jGUqcɪ 58i# knNz 6TnꂗS?ZC/ |_Z^̫p-fS'mH ;bHaN351leLjVpdE)ֵ%[ÛfnE_jf,+7-26eLqϼ",+bް0qM#iSw( +0Iv`A6`u1@bYBT)ZMPI WjnjsתhdS0QfDpn 9=q^04xF 9'Uе Bڭ)ΠL.PXQFNіc0w& Ӄkq=1S4isi̿LpZ<9A}]7Wi$D@<6݃lXrӡ[dh܉陻Va}j!l|0%ݎ0MԒ+eI(GN&8Lhq#rP5%{Yrn߰R c9(AjIxL3" oPvuZҎM!DRDslJ0rtw3*[.16Nw-6woD3l8wOsA˜xQUS~[vؽ[hgb73.3 ,7kv0t!$Ѩ#] ExnJ1ԺKZo}g B,~b3﷥U~l퇥pJ^_ l>3{7_BEt;ä?~alWo;4 ;~p Fhl,٦gD1fI!D6,S%};,U%-؎%^3>e  |f֊i}:KQfYG+1ro}2Zx~8_'VOgے&/[>Xl~nI=ڊj-R0uK`I8}א=_ŒΚR8j;YpWX ocċ:Zs[h*tb>c b~!)b9Mt T1F AG_lht(CcuG 뎱ɘƄ mUW,NUU:'5*EbPC|Kᴳ%YY1.dh#DS7{N;[R9SpTB'D)qƕQz#a 98j:fXdu[ $bfRvZn|KbfB(&w .D`:-7N5/M~Q y\acN˂ON϶!1F64;6U҈AJNT$ܤdϷ%RH%ӃmVx$ Z)   0H҅KǪmĴ KWPƐz½MZ9^5Q"x59^8Maz0T0$S_;nU40[iag,+a+0HYi^v[/"j7k?.3΍ϰLYWɷjbioWP2-YC>,iz)JЂ5xֶgAgM`+:bTtd8pdE7t;z26 h 蔈wt/p>p1>㲮TI '?MಾGޗ6~(`KFb;֊ptrIw%˧W8׸"D=K.Y $I Kpi s?=i_Ċ8 3]RTiU%PtXn\rH`)p 6rIquYp!Xa5+burIKt$$<7 hEa  $}2c%w eŵր0K-jKjseܺg3Z/y~ &26ĚZ'? ǥݜzopɦ'X_ vN:3k.6 |_M]@hYͼa$)9::#t}W9ዕa?m?}1eO= {ߐE4a:u{v́ ~J*P_3.8𕎗<֡gH t-;DaSgLm V﷥DЋH_ n8H_H- +=k+"JniI~[je?o++h6miL++~#u`^|ҘP(]~8~CRCzq-Y&:pjZn#H <}0Ȩ%`/1-ZyBTM?t9zF Ƶ@Vŕo +ƫ/,>3xC然Pp^KJ*[[hx1PJ-@uӺuqJ/A9㧹S" z&A`ME+^[pJ:1a%Yzz{Ֆ3p°H2wg@)p(aiX%_X^q Qu,fPV۰s b  G,jU:dMz*ղVNlHp#k#i7Hҿx4MT\4F0x7nc=r7G*~z6tvD[sʀvbMaƄtEc1Gᴲ֖,$vA\rjwRE`QM}mI{sbX_Q_A'%: (tnfs3aM!TZPyr彋\̘3,wd(2O ;TyV2ĜY Y)y3B<97\1eV:@HyG#CC[RSތe)"ޗ nC(@;Z֡9-QKt9HT ^ll΅ 0Vz{ F"-C ̓dǜL꾩" %4?0Wԣ6H:H1vpeZ÷稅1\p$}m7K)c %ͤk= &l)vPVgŻh+Yk 7Tك=pưXu=ڐDP^.)# Yjg!"CrZu_M'`6p^,4WFTW>{^#Vn23n=KnBo@P$pE i荚j+.!ѣ`(RjV< Q0 O (Hbf:s(LT}Qf W~>M0}p"=q2iT­uQM$9.iDh/Hɡ tX_>AD, T c,g~V!J(w[׬fȵ4׉uFH^tI76>soCq`C'"n<L"_H\ t[5|ň6qI IC@L纴BmT/$oc7OJJ[/|Q",;kuW*&XapcB~pkh\E.\(5RtwAgC(Pʌu4! ,-#rmcw")rԚӯ#~viڲJ4J`&7]tFŹEz >ү7ٽLu)%ofV#]ᮉ"s_=S˒{hEe‰2W*7(|n< 0,XeGF )U0j2 U8Raoጒ)M"h OG| nѱ5Rr1G$$14xt}]j}rdF<6kkv^sc=:l%tjl#LGVT=4us0j?7z&`LXƽ{PbF?FqplS_7-iTߛ$% E?`i?UC%@hFm[_PL7RtݨSmGnm MMW'nAj'ʀb:}ݦKN5ϻoCnCX7nmJYˬcH'VsÑ 7@$R+Xc$l{:D;+{ǐ(D Beigax$Rqp6 EtV,h0 uhE85;E9fYOxӗJĶpᅰW;Ykq4å DuH1r( 9~ƅpn3!2!0#G<ױ;^tS}ݦ:|ҋ#  2߰D;Z)bv&+r༐4 /bCv;,6D+dRC-%sC=?[{k6sMb2{-@51>$8J+6vx+=G+EEBeZFuj^Ts\ê)RIU1G-K)VO,쪤bꕲr!͔$=*QlH$bwk~\?- c 1#ڔD#W|Nڻ*bڶ园H&LS%OI%ya:DD|/iĚѾîEk):Bj3;+ /g ?ˍ!5zijvoXK%|.H܌4>墣n:RZ9n:kSڇp"|΄eMlː" mMRZ[ dvY-EμB[g(gPǖͼ-2Ll8#>W3:שnSg>^D+ZPf<1XvY5+}i,S1yJU%JtKšBTQkRz(DPq bY {`|Z 6_c\VɵZ]k\d R I oAty|ƅ”TFVA pivՖ0vQ<%$dR˭wY鵵+Hxa#Yَj,l:ydT=M_CoEmb H̿z' &OZXfy(C_-z9Fve fP|Rδ-W!yP\ʻ^ٌ83YP rζ=sI.|^z0 xI}$ūuUc./ᣋ>1+Ru~FӤa$H$L'܀ Izpև>F#:++,ʅ9j/Dƌ/ 81JrJȤ—(3>FDskmQ "I Rs[a@dD+( aƅU0&e m.$0$(Ii}Bfی*!! B(a^"t,y^01ȭF;zU<7*.Hs&dsYn2]rE0JF T 9mi0}.m|d>7}H ;!"Pr }FB4Zamp9_GڔsXY JB f\B+]1jD Zm-xϱB7׫ ?UU`:HX o`*6 ]VoWwFÇy!rV̡6A?Cd́GAs-W?+92a >5D:0~Pڢ;q6>\+[/t-/ {@g;3"2"-zC[k9ϠX؜aI-q> /]_/ҟv S]9״t_>ݮ*EȋVOOf+ν ?b%py懏1ue_&|ϭg&'mPf>&>8;ZqXb.g\g@"$^klll7uSs {a6l/*wwT5$ZR'f L$[dtNJe}qr1h9oZcޮhk\uxl<.ҢEuCEZ/~0 ҴV. R`e#vwzMy?&hTpy4s&]YOCk &&`| !nSYnSD{ QZ HapU߃ m|'A*!yG?8ZlGv[f;0f>XMTsl{cpSR~l)+f'\w֍aW;3l;cuQ]?/mܪiG󰈻/>]aCկKn}Y#}eۗZ}Ҁc|y_=ru23vsjTW9Y"6Wӿ7[#li]QMagjDBqeS];Uލ&lcBW6+ͻh]Cn\6iNʀb:}ݦzNAXh˦`~ȸ,k7m>Xw y`v{Szٺw.6p kaiA՝'{9} )ykdWCf abW-XSeLT@5R`A7UMVȅ=L4S(#ѵ{}aH)<_(ђjݥ86F[#6~v~.{Ϋ+|\V@2֙e%:fH6kVW-#숚[VkaZ/~j0aBC|k2>5چQsq]րj7KlASC黨g H^/ip>rkK@J]`{6K?$1L( ؟E xV:oČy<27ʑdB0όmYZv]Blzϲ2-hgj{_eqk Y|70,pCAnl0`w=#-߯ؒ,f[8 27Ȫ Ԁbzܹ()c9&U>Dڻ0V4 D< :C`ZSZ9oNR(!g@ SNf}2@uUS"_DӂTwK:fNٗ-=(A/sm %DϮaXB 3 GC;J hSPb4 ut{!m_ $=a_AxB8.vl{նj|ԶzrX):$e%쬞]g'K Ѵ~*>S`808Ҩф6DL1ܗ_m?c` p#&vu?,N}['@m1's1HF# %l ȄJdBEl@az$r*N48M=ƅ:@ަODo )3@M(__}~h5E3ZbМyh6iaw<ى~{J"Cmz5XlWYg5+pZYyю2D'Ω_gy3D1:zLO㵨`:dL_ $EfU5847v){o.,W7_CPNe7.~${Am BN_؍oUXV0#@a/Y'CP[tqTV6uM\_K0iӈZU{IS^sp-9X8Ovz/[){O(!I5{C— >8Ðoo$gnk7BIOu4U# EkBS9V=Sp@EF JvA{u\F9' JZX"~^PCf@=[VهboJ_t>[Goe5E~/߲Ħc s"@0=Huܪ ٽޙZsmx#hif A?W7w]@Izi{fvh x"|Q(HSŔm m ڃeZK[ShT$XDid JZfA eHy$p,2#*Ej(qDp(=l߬Z9XZ)4BoTleb4ɇ_/ j$EF1JPG N JHCx?gbLlbN :dD8y1GtF/aH_A432siM80gQL4xWKgDž43 L\]$E ;S$KLrb5w{M1yq _ck@ Љi $Fa$% NMIANvK-[89P2{ 6ϓs(^TpŨH≪>OP # \vĻC%CzPJw[ &{p' 0>Re^h&hΐc}.i}F"d[ߣxwYp0-+*%JgA3:٬֣_ߟ__{uwe< >adVyюo?8\C{/t!st(Uz eP֫ JJR⡊+*MmoC-_{3iDg1?;e'ׯd˺MO=rl8ȶ}g+Jg3\`4 J+KD.Ω˯%Ubg@СKD /¡[C3ESnƋ5_B50vw.:E#w_a/ts)΄|(B^5 `#C2ӝs=c:^ #^WO8:6{9{S"3\f ~j&nⱅ0SKf%#(7z‡gTn{N^Cw`Y|KVg=ia"ŭ[!oP_M\Oߋej~Bxm{?߄e- w'8FO{7!6yCTVb4G+i>*WY7nE6O?z4p ޭT)S6alktޭ%ѻ a!߸ٔR'aO)$F撃`=ۋ㸈;*x+㍐9B'(JF-OP¦^{;l{>+ܗfe[E|G$<7*obqxjSk'YX.qp5nahd_{L p&A K1VxceW;w;|#q~x e벪oBVb'h{(m7giSd_SӂR$5ԆWW}NmmO^װ[՘fPMW$">|gZ؛'| [}3 %by'x3d2ikգ6ƙ_2 X2c;U3@{l)ݼiV{f{Ԋǯ1dO1kL>]m,UcI&A 5;yn7p@>*FJtOK]l%![sbr˥L_)˥\}j0s$IP $AW g>4kXMh ƀ(aMZgԊ7zIfc0PY&T=fZJRҀuP҂$B8J k4˸ VxK:B%rBi.t:{M2lwލj+q%WAY=ո-3,tOMZ5MU g+g3Up:m:~2cE*Õ8nɕƠ"J<U{qTs-$Ԅ),vA zR7\^*FO2L9ש6t9ᔓѪޚ̕`)9cjA`Huh 6BOfZHY0RW'+"mP{q 1 Լ&! SH[Lt6Vh 60ѾӨltY_)m@sێxP(;ڬ=P$}6Zn^H2XBL.u^=0K;5>Ny8yE Yd{ =T%9K8\)58G.yB4Թsbuz۩LdE !˜J/gEF+uGJ$XͥdK=~ZV2^qGHek*t (uN:9>7hA5 t𵣆{\`bpf$du!V[)do*$픬}uwQn6<j?lS˻{`py#GsmqπrZGτK64r ɮ3##X}\1ff[ /H?V]aYWub5.Iiuc-T{qVAU+wAKa[7ӑpt7 7N|+ Dࢆ7gqK N%ڪڽI-[;l*-jK2nyc N0xOodP$)@3-r Q)XyT! й6E.|hZ1s3LU'T*2ײJ/l)+$$rKEE>jBmkGI~їXVﲺhpE69V3ۿ{L:cKX~e=>]|*돾s>ܺgEV~!}yyQ]7NoQlX}Ň_(( p֮ӻnƃ|u?#i;7kpL6DS \C`ˋyAkPc}9@%'X1 ZgWnI, H+b ha, μXXQ\*.-3 s..[9{AmY"~Yzݠrr+>w?_޺xfZe}ke.dqr޽wU>Gl^ٷkq[AWkIX+*&_ݭD@:Y4GvL]#ս`3nDg}[q*NSj!m ztZjݜyZ(I '+tZPAjL0ROc=&a7(:#Z hd@,J< 3nZ5ZiS"4VsKGYܹX0BQjL\p 9Y/joـ܇Kvo.vZgu3ձkeI68K?DNB&xk !lmTfk:ŵ}%E^KIc8DMs磥Vh¸}H^b*x?ůmw2B3- (?;-b*$gm`H|^>}nk3ܝru*-PKyT'oDByp{Ŕ-\ΰJgXa PE菇ջ8~'"^cYv/p~P$=8yAFأ ūn"B#$\*ґr^/s_ks$(I5.feA0\?}@ 4rR>ơҰ.#Rځ$ L Ѧ5ijSUb)XBbYٺ?Ue *J2-s+2Y)~$o Ր5_6Z%q/Uz d QbicrJE)3]iL5W/(SmxȁY6?.6?..ߐ$׀.5i=,%Q ΅OVo\3Ģf`eY)d%m&LgQSlS<[S(aE.TvSRO+|B5V9usi|[ᲥbH,'.46nd5cUUЙֽ310^m4l"}-u/V9k'5#QѪDLp5>CJDW2ħ'6 ¶M5UC fR(i`x5=|pMӣKq7( !]c9vG,y&1ǂs-bjo?7:o0FkPOcFIJ@5(:`؛hjq @ݎ{˽V_*%mKWQ+:J{czPmIQ#5F ԑelzo;SMJ+ejCTkW *-?6O6ܬ,ۋk] (rXȻظ=s/~|>.?}ǧ~oMyW9w'L֮ҺJŧⅫa˶Ղܞ=z4djg#ؾor}S6|*KH~~fݸMh Du#ƺMwyͺZ!43W^:E8{u$` Du#ƺMEiǶnNA (dFƞśoagf7) /i,@a-U"LKo7YhaZZSM⬥'aiZލSj!m5sSRzT{[[d=PZˎ.[ABޝn>͝S]䑧\$ZN/>= G$M(S!F"z%S eIF*prS.iQO۳qC1 EUY]5K%sd>2 kTɟ~ɷerzVdh}-EQ7Ϋ5|C8iJ%?es$_2dXbO,[Ebv| d;]s]U61])dJ^F^(6ՂE njȆ yβ6AL#>˨)VG.zM'ِpVOdFm6[+u}$xX!edʼLIeB*5CI{a%ZQfą3HFE):4~ np)LB%*!*J[f(aJYf&%KȤ1хؗ+iRd_vH2Q%npC=0|*ڀzD ݰ4~mYN@a̫X&J)Q]ۗ𩠲Gσn1OocM~láyk2ɃU}<|ԝ!&_,9ZЎf3/O ]=-" }Ց`H3?LM y?-5иFS)f@ E/ة¹Ma\c;q AҠfD!W+?/ѓ^҈)$lLE Y^dVTh5/N;e5AVe6%Uբkѷ!jIg9F)6Մgdk)j)'JJ9Yv"/phT_w6$5 3X#nz񛚭W&ڨF@'eWs@c1sڢX"A+.P 2 LMPJlE(WXei%ǺƅJ^l!{"z:<D\FաNkvQ?ahv> Pwj]|בg-{,^ʀKvQit9@qyo4 &Az8*D38DڽJs&sUiILciY XT9̌-ipr?8J+/m(b^"}xhN\URh*ΫqYd EF,RT 0J)aPɴ"+Q"i.q.s6^[+FHmT٨^-ľ'eW<„ƾyB@].uKtmq^5UEEPb,QV)Zk֕O>cϏ] }@GX/Po!ư~| ˯?~UY z!ptzpO\|ݗΨ7n;6aI,o_|x,⒬:Pq}ˢg"J$./n6G7ز'cx"~KݰwNoݠ- D!I43Q V;0NA/աvՃļQ;}^߿}9dbGsyAڞ 0$(H-{N(.$vm?ANsm=j&5j& D9kH9ZJsd}fwgfg5xq&!zNIzxqRnQ)fǛ8& 5b{w~ztUM Q5BFjQkiV*PUfk gEDz#waJ4?짇n]9f@)m:])`1Z^cnΧPJ]V!2\-vptJNHpNThqp7\LG;Lȑ>lب&0Xh*=`bq8 ϳ\*8:UxǛk4+7PR^׆6 IGV<먝k)wM+ K1!4tFQH&l(> b)/ڃu4'fxSh*(P:w>Ѡe8aLlDLLGtEԒ3sɭC#̸2SHha"r.5JK$V.8J YL)}ՔL!IHEU*KŤD)Bƌ{br9{08jen6F8DLYQ?k"?wb.hAۃFp8I=\+e #b$$hr@\$4j>]`O3S4ZOc!ZI."HMZHbUa6|#&yqHHnEƹ"S?<Ap0}bV91rcѰ aDW0Jh`( HS1XI;i{}I0,'ĶU)}HO<; ALdę|L Wji&UfF)TrO{2֠xmDtTsx"#8@@{Y~VyfوÝh osL{&"ϐZ<m2+”5UٖHimAY%)&ꩫHt<ϗxB`m!&03B#zܨ #e:kC`Wnݏ" x,<  90':/ XpE =|"<́}$X0h/xny02.o1ҁfjy擞?;"MWl$n]wmJOJN@n:5<\ܴ2x0eV$se7?艙0\N3%>}6-angw*[wN'l/3t n]i~ L 1`0xeRhҍm;XqĺDcZ.јWVoel- mwɅ6WObThFupPFb]G!oH-RK.^rԐYˆ`roC}Q¬@P =%w`1nyC`u;u~YQꬵ' &Ŧ$ţnUF„ޭ+)Mt1ghޭDև|pdSiuWLݺbb:]ĻM:DݺיIn}X762 9qRqVh#Hzd&Jz4(BûO,(&+6C C"B0S^9Shh"Rvj"-J iTR,:N g봔Y1L-gi$dF@cFc&dFE -E;E>NL6*Ӳ}"˚D̰vNJ vD╲Ȅx%.Y3ulxL0#E/z?]܃֎pKOU84k<-#篸R ؄A{o/.?ϪD>\E݀I3f=@MT =2d6 i'iD=uдц޼a{VRI>zi/[aKwASaۺG-P%OPV!RǚjzfLNKzxf9ۛ|rѵ_^R7iT($lF H{&}0p4#{$6VyJeM!ACM^>b) .ChY]V -)R1]R|۝vE4#ܹI{Q|cZ $ +3Bc"oK7յWk)/6s)qkhu q]1HI>&q#tbAab~\MM#x7лu twQ;Rlahֽl^wB>v) #κwcӣ;nӭ"N/v\$95vB>lJ%ڔDIDϙi{e7sW]!fڝ^riWKlYd; p+T;so>8%/O f+猵`pwDݕr}S]y6QprMe%R7SJNQ}ʝRN)#waJPLg#{>JUҾ ;*ѤRtjՃJ4dJՋJ@(e>Ѥ<8jߣbnl${3q}OԾ+D˽"t{5#i[o{ a}b"HR~vq\l#e+q~5$t5ݷ=7j_{}Z,2g&[NW=|an{m9x~$*e5jNl<6kn`UV0ެ>jj/FũrB]D lsg<.?q\C0dϗou/QsSiя5 FC۷/.|[v$[zYfw*bSr{K9GcⰩ|j=_Tw1r9┲X|ai3K'9/Y{]n%05L#/o;Au0V^O_v8l Ž[wϑqB s|sq 1o=(2ms46[cL5Y7F8uE ·+5CE*xj-&]v_R5!>c=c};C/6kOz9:ᤲ֏Nk-MԅxB@RC~U)'Uz}٫ǯDt9"T5z?/T_QPGMtw0z *őEFwMq!g m4R|V[޸gb_K.{?ܾ|t2y|o51/G 5cR{+ u$S>e/D0E^zG79 ɬE,s!KkmKI5I"ͨ겠}17Wi3-^w\>^_]Qf$GQiZ[jO.,=.bgŗo"S+=Y}b93{J5l;Y%2I85>+F`}-A9jjPKDO I|,Yd`٦_gJ)Ma[5^H.KVets?]?{WȍJ}^gXh{v^LrJR}:RW)S:WaR3 ~,tšGp(rspgf{"LTs }|I\}.tͩ%OhwU@~W߃Z$^?7W?ǓpNgz?)>œ.fכcRdݏ|R^:{^|S}vW?8;<(,Z}!ۧ?_CK){ףp=)j7{LD)^Cb VTo=q_Ft~хr5S0 x؏}tӯa|(wh[wI+wU<Ǜ!i9tʃۑͦ5n~0/W?d ~%Rr`7)IÇ7|-Iڏ9̗ٷ^JZp+SPϫD~\/ |N7oJ]ҭnOf?*yj4z52fNBO}N?OoOBI n殿oBf OTB8:wnzRfk$P~>%D i#͸)- 9qd.ؼiCE>/ihZ48{_DE|{ſ&`8|8F\)&5K;ɚ\EMW/h^T|͜o_ֺ-UKG hZ&\Fms@UMUr)xs@}<%F<>D絏Vc`;f. % fکgGkө.b,'*cĮN$ZdtsNj_R GN.7 (k&D||9VwWT<_\nkW *b_"Qbɛhf8Yrw-4][ͼ$F3oЁ NB3"!Za-ZxxJMlLp"iC^dzIYkqnﯞs} S[sTlpRޮ @~N4lcw"?16D VJrIޚΈ'NטhGv]IZ`[U>E#Gϼ]`r:ap X`4C\ i(pYh;vY=R?hl8Js`B e) uŀf+]vҎx#$WNzQN;cLu(Ob7]:;]txD5Ç TxCj0Dgs?d`󼽣5A{*U!'מ g`=ɩX/;ԑXH-aA '<24%!ʬ$W1P/+:qz[}pl1V#1A(WaA%2`Sä@q:QC2V Wm\f1efoLPFy&#Tǐ\K@q`5rȊ ,A{SdN53cs* jET]9-Aſ/fc͎쌰I2f@E8`9֍Y~N5BjkN-4 Y'ŵJjwmQrUzK5=*R(i S r-XTܲ_늹'0!+rt'0):]x5Mrizz4=x҃/?1甀rDOډGf2'zoQ){V6.::_ʖLHEӮjg Yk$ GXLz0 UZX%8S=*DtXȤ`R1 DS]q B9-h6ʫo'tJɨ$" zN5u<0,RV3š4j#{.NWqd]X #EJJ!NC{ z\71+. O87sZ%53-j_%TJ&Ya”lvPwSB(Ԝ~4a빋ZrQ66Jq7Ȕ4#SR.uzX/t)nQH{Bg=J3~b|ΤcNj :~ 5 ekz4 95rr&ʽAm,[l\DUq]eIFG&eP;ku4k40]װtD: Z N3Nk:Q+ʑPhX߈`UDNPA#8rjĵ)EI:=l9>&{J~&h68v.P3ej-X(@roH3?Ն&àL ԒbGa Y1LYaő()դf(@jo=$Y/.|'Ôcq<|I]DW|U=[%,E5d@̸]+3GqP5?l~BȢݰk,a]f^M Ԉ ge5aZɒf1vۧM I&).7kއ 2:Nha [G< f5HéaˈO☯Ձ,bשT+sVJaV& ].x¾;e߾l7G ZδܷRd"r;˂A i)_he7O/,OO[a <HxMHRK)TnTY9P~,|%Ʋ&/d0i4[ w}X~DJzB!u^Ɖ>\֝y(I Ͳi+zu2:^>Iډ }|%go~b')u0#q d@zo >Z~%;?7Tb}2f߱2u"Ƌ>7y! 3HA9̢hR>SI6plqe"V 2,MHJ&4H9VgnBwϤʸ5u"Cƃif c1#*qLPHQ@,E+H 1]LHŕ3P> ?躹7"ڵR, $!Mi_&`0ܤ | .?mU%&躇a\6ҳ|8 Qvicݨp6j~XM?\?WzSy/³aUm[Eq7b`;3I*"574Pncr{]'iGgΠ,p#e'+.NKQ'ؖ-dy(@NZtrBQg0-Hfb]Y#.li4:עfV:0.ݏŕyF ™6 {MZZ}7><.͖:ױ6T,c,NSæ9?o{N[CuBQifZ S3uRĨ"a? +23fX`݅|_D?iWIT<dB+wFH4NQ Q&u?9̍3c q rr0Cȇh$8Oij@r>A&gs~qg!O qV8 `TCt \d4" G8pׄJ77mAIgƍs`Wbk`}P+6Q m@Oqo/6ʣ퉑כmg.Dr 'LdbHe#K袍H8Eoltn[S1ku;8 {v)QWW/DW6^6$8C>7)KUtZ% a@H{"rR~eƩS.6%_ 9ϵ_qеtQ[8yW h$\)Yt1;mˍmɞ#TH+s 2O^坧%23VZ/L#:YEr.n~ɕCHcʚH_aGJudee1v>m7'cN"JDJ[ųI$nBu#γ.@\pǧ߰&=xGSްkCDh >G(/WL0D4ZE w}m1p#g՗}F4vdt(7f?d[7~\*n$ Ya4=,&FȄ4Nm7p;+>iLz<D<ŕA!p )*="<}[]Z}{M?%*" >Tf9@57ly0p/I%yt4Czݾ#>Od\ՠv!spbN 뽷 i! !svsVwomC9,u԰. _"g>~Fjޛmwjep]{  {Rr2U7J #4}e. jnwUW%ˆi:Rd'1hbAJ}75+I)"_i{a$'x~.?YT n.O|OlKrǂoܠ5X;8`"_3@B VZ!+hE\-m7qHxXiw\ C%)H X-|ZjeHlEd;$b1],:&Vj)Ͳ5/hR!m;m0-inUE?mKhÂ!·/"ڂ9y9V[9ǩ {x˾_tMі-!NLtdj$e`hCJs,YdVE"V6M$4Sljz`0y ۞7'zYL.L :ޒQj)6VIQ .<0 1zxM 9X/VC1Q*,.g騧&< zDFZdBAI=0*Xt5ڔsa$pLAD V!b[CdPkWGЏN,Y{<d̠:TRzrut' a-pF/.ލ clbN]rm16Y…LZ!Y=!ܺ{Ğ^c;dioVqMdd0Sj^>p!r_~F;0RkFQ0{<#G ɬ\pHTk=8RA J;L6R.APb}da:\.F0@yhdRNi+nuUhcœ5r-bT-l3 9={ܱr҈kobqFZ%/j9arL# :*?65lJrWu&rۅj#WFOvۻ?H :xOߏP yX6°3lOf-iM2&eQUk=y(<)&\ѡ˥k-{h%o>P%VWU DU}0ء*1p`'&R tŻͺAOMZ։&5ɗ PGt}$u9\ǭbY3??jSߺ5Ŕ/ƒۧWw=_.61oqQ{jIzC}p-:zD #a1 2kyŰJ=vRiXv!V8-ʼ}YRZpOA#M:L.|wNur>F)!z{YVqڰ:..ιP:i?¶Ѩ5IcdW)6)rxkN=ĨM hx2dXYՊٻ;f6vp^ԙڊf!u8#bL)qKrx5m@w88Mհp^SU,A2[=.D`%\Zm!5dV҃] J>L{E)ي1N.vމXIo !.SfV)ETVATi"R}z⾛S$iZBz¿ɴs:mdkۿ1iD}Iv5ґ(s]8ѩ0R{ciEA%[4>o-K&*9McD5rT|~JWU}NgSګW0jE,9ľͥueJplVsWlPa!6&Sϸ; JUP,z0Q8T&isgyM/~eA|rT^qM~3&*כiwvhrF*9=){rJ[ѳZ~Lղ=x\1@ hB=3`t$ˆ(MȎSYPA{ٚ]6ZO}Π‹$vL{poV3IA 0FH&D(ZO(Γm[NQ:DHsxOCx܉q`u5^ -ʀWYAB-WEG+V?wm5c|hq~RhTҪқ52- jR=^uգ zӗ/KQkt2 nFj.X*2h-RҏggqMT {#ia?$فQH<ЈQEmF)8P3wBkSImkCϑ!P6bVޓ:T07ظȱ1tF4ζhy|-IvcxѮE{kD8!&i:Y"Mj9CPSlq՘o|;,w`)R;5 Ac K9 ;Qo#xM |~z,~ ￷YxbEoWݴv7JzOgܵ(gHYnJo;vu[h9U/ uW\??r(V{wm. w_W?wGe2_,鿿Nw] W?GB_t?3Lf9|v+튏?߹Z)V [W+_om%MAHq;<>:.O=Wxu}Xb4=[t3Y)Yko'\݄,Tҟ!ڈt YdTɟ+IC miGg.ml 58̨F8KO:6V&0B11FSuYIt)fbv\P:F$k6l#tGcvHz:DfF `0fE^-Z4idk>ԭ(QUEQjۆHUq*2.q§/je>;X x>?05+[b_{_ec"BA6 xϳ0kw h'~[O%6l1*7k#]Zᚴ\#'!xU2o\k,f_QɴguɍbPHToH)Rˢ4>͙uB^n8mpHZ (.ձh%`6bE0 3QMȁ4\Γ+ ZީПs-[`i ߭dl[G+muc}% J8IULRHs~6NIf~'aqD5f:[3,AON*#K$VA*X0l-qwr>J@H)O`s`IWIiӮ\kdtjkS%2elZn+];P9c>FcNǢtL : \*xvFg߱4oǿ VV|e@ %;áV ?~jlUc4PA.mCyx.6qJ.W;bjci?l-VtLL)d0trd0;&_.8ڶ.C;Vt24_}bk?Z%͐{T+$Fb1P8 *0t޳/4N|߮0 5'Zu+x# ٓ^۫AU [ƥ/f:s |V=P=@ <,w/!;RN~?i>^ur}Jy#|j4YU2yQƖDcՓBa 6j4p $9sޤkmika?Fi#ԞR*s,י]2`Ɉn h;x/ gk;{-ZA:t(Sg8Ml i}s ѺdIYx~pSNLr-F9]F(ih aAjº_+g9T%\t̡6Hj_ZvIc%es Ȁm Sw^+YP~:ԒRNs;uR핱¹6y)4%k3SީP-Jx-a7e߄r}vYUNv5GFd<ͤ[0t3<1Cd`\&! VRW K2R ճ-B!)DG7h»JB K^r:H3ϗ 6Ok*2V 2w 4˽~i L,c 2&GKtRmO!=<;~ RAY ,S_n ~6~ Κe=ʂ3CzvIIJȁ6sƛ,X..FgL* lؐz'_&bVv5%GV(Qi}͢Έœ} EBnٟ[#r;\n_nGG]}?/'k ȑ sh[oE8 =*Pl9S; 4n^(m`3!9e뺔{bVeo ꍶȬ:~BH[OT; E ܦxETl_$Cm9) $Sճ. VeeK{;>Q[>F$CcR $>3 +Uj#{ESe>)#ì3FxIɇڄDlg`tfK1 ! WTtƽ ҆v ,'V޲Drלm6ľj-BL1Nfb7nNg& Tөe7ULq|_CEI;:=(kwmPtW#mGZr} ryoGyY5H_?&ףwz<Ӌ&JFUvb 9L#o_7K魥-ˍ{Z=;w\8/^@`͙9/(Gp^#O*_dT&(b̂RҒ{tpzrw577]С7s4^)Qr4^xc9lr. 98-4_rӑ?TvhdޫYbG;F~;ՊĿD~4-Fi=t>0쀛ĒDF*:FʓgY i_D(ݾ"-0M*D"yWQN]ؘ>  Nl$h !#whn\&tQ!:m$+$ Z*J>9\WjE-Tgv(ΰ*@:A.1hɺ.\h)1׉\ų[ C.=;v3I'fo䯰%2CM3 фnsYpҘ"CmwhU Rk|dbf"٤[dߓ/ZdQP0~ ÅKNnV>{I( DnBe>H4 te 4y=Iѧ_N bvnRqzL)MWyhKńe_Ѯ/L&`u`twE꤅Q)mF!uMT]]F.ϱIUеέ E^DU,t7}AEq$A19/4,wy9w r|4/f.`.`Ȟ;wilzx=D .^75Be,[Kgh!r|)3AfqJbH  /Ng<S#gpםKrS賲* r ' HfKvH)w!A شZP{ȃז͇eX`~8gNPb=ӗ_J C1M'(S,㘄˳FBq+ Ry*-PK#,th9W>bW`o978 COJ6'pX :xRI's7ܫP6+M|pɫAegsh\/H Jh Xf%-3.+C;U$:cFdGD ku+2p`i OAF金ދ@q*( QŤRgoME1x (\ɫ٦䧮~ZV !ڲpk:ҥ1-Ŭ+=?&7s33@dEhfX2SYz(с3s0caVfV&Rk: >/r4G!dڏx}hLg.̨j M)Mt1ְ6xBC*dM^Q6hr܎CYpu:nkޜJHl78&v5Kϛ+G)ˁ՛.N _ ZmYL1vuZކrY)ڟzF߳;U٠0BdJgҚn컖]0# 09r)bu \c&"l-ˬJy0&$OQqknT|G7d"Vvz l&;<ۤ!i:$nmC,.[)-2̜,!`d0ʾڡZ B*JQtD?A;BKd oښmDu5kTETjrpݲgFg;t+_j%QiY^|W}nh^\ϩ4m?XQ,Uػ6$W ?,fR+22흇}[, Ԣ$[ZvY~v ~N0!5_V&O0I'YYFƎCTbN]"ZvTV2$% sR3b+1e4@/۟:! W 'DL;$W`] =b.Tg~Nl(wyb,(ԕ,ZQ nl|1gJ'_{մ5k㍘NHj*Γ*t1[Ԇ ¥L&d܇(v67YnӇ;~#kĀ6g 4VtdإPUG,4 U` -{=Dv@- P3)ό)O(e.`I˩xLFAXfGjR@ZnFR=k)\!Y C\ S-fPو-Co\NbS£wUd}<>trP vX_**vp(Foџhe" :F-6ǐZbAFsow+ 5鏄1q6\1;OIhU0`lrae,iťEEv}SQ| /fS.-!%\Qӓ sVHDPOe$G-P]Ku %=a1m hv.#w/OD * nu&h* r]dEST 5Z}eH օjlַN 퀓ZzIo"v(eh|QB0`"&wm1wsr`zG5S/Sq <̘yS ˸l˔I^ !L#OA}3K5Q$-`~F>>ٟ+MB ts:٫NXS={5fvGJew=,kWDt}FcW2{5K?lqF7d=Jdkf"d^Rg7gItu3Z9Jfv+Aü , Vל#^rb4gW#egCf=\`,mo 6tS޹mX9Oѽp>l{02Qbd["+(,÷؍ Ui4 })P4BX љ1gz?FlN9(14B&s,oh䊴s}qMR,ߧ/;+B,EAgdyF۔ HmK\ V^$Rlamge9kO=[Hk=J0D!I^8>R"J`P:[ ylsnv?]{%{vC+qܸ}0|K{eo҂rS{v84\f_)TmZv^WS^VECȰC4#i2(j#Toت<SFJYȤdKW< ɑnixv}$9;!*^#lYg]S((.CJ* K)% Pi!fdLIr3>/ä}[|f:P5I)ō`X_=,t8'5{}u[)\A 8Ǯ[6yo\;tTݽNP 1fa~E+]ZX 1(N]q#:!*m|9wkt11r.w-"Zo";-|mv%6#}V:8ǂrKV2057{\|_)|bN0861 ZbK o=o4P/D]ʿELtB07.lDq֛YLd/w,Y3[F=6,Cpph'Nܞ $ߡ(97cjxѵBI*/B.)F|rRcj|;׏V&Qnc`(90 3'H~@ ) JS^VZxkl̲ɜu΍;n _Fsc^|P&fωsצE d>l(ϟMC& zsKlЏqFM3E +;oRf'6I"|2Ӗ4jruq]B&Ae)"yBK,9qi3BoL7ҵpDro53)3)~Ɲ5@^4%JO??W;cл?ݬQ3Ϧi Ⱥ]@I g`KӚDfԟ?iqi3Us׿zSS(<$섅C 8H|iD <>&qN7艌d?^(=b/nߵ"-$er9:4%KnRg^F/~1/:n[uN{vKtώ1eFt@EZpF BQދgcE&UFd6BV妅wR G#>C :Oc؀cq5,5v;( }ayn9YKFI .Gh|VZ)MRU2yi֘$HlP2pd&FU Zu :]ۺYry,ݓYڊiH),[' hza=]C9@AEOx*0e?ErP=ٮS Õ:(V eaHR2-1 f"QA.jD8IbvIZ[_r2 x0Y2xFWY=gU~<JdM)6X4)*QGNj@ȉk#Mút:W&.+'@pu0nb0cIW ~ܚOjӋ 1&1yOf4'|˒_][ѧZg=8޸-ܧP$QFQT!"ߐ.% Pi!!cJ]<o ʇ[*8߉&ms8@ 7nq^i1B03׳ǖܼ.!z^ć CS>Asʾĩ>fIޓWH$yQx6ZT<+EuR0r=Q)iCDˠT u(2aIUG5*uO ]B$$ge\ȕˆ90)ӄTza+<k зfH$7,z¦+]Ys[;+~g=Տ5o=53ﷸھHNr*^d-g崫ؖ# B `EЩ#Ԛ$ 84EHp9{@ݶۻ8PƚY~7˾K |BB4V` ٚ5r䋳:*29XetRzK [h~|-laxYm>M.ĀXOkdhX{()/:I;]!wQ?x WL0KaVZ0E**oڝ\87i>F̍!+mӂ X󵔪+n˜[WGKJ9ޠGJ-ȍN{E*/4)U ;XEsA7V0K"H'!кlHVK3$t1iȪȭ ]{ü^ n3Z5e2kfʙCN d !Z\0ɘT"}1\?, zܜ8 nc d,cLgL , vB*y NA>5-'[߷V뭅2w9e=v\>mR夂6Y$lBUpA$UR !H^\!u. 4"B>MPEk|K(ty !U!r% 6ֈR:w~𜩺YUSWApaOj46`k]tb#H}f/` s!d92s#uYV!5 ujh(+U[_̥>ķoA.PvV2C۷ ½|Y!,_ *C4CtubPe)nxķ')+oOQ+tb҇e]-wDvLT5YEX#WO33!$'{o=֨tn{{ʸiAТ FHIϯȫ{S%QdN7:BĎ!htȜБX[X4ҁD:-ysf;lc|l s,A > %4f D-}orjzoR& qsw{ED6OR2]OdSg:} 3g dYA :-O;P!Ig5 W9]!f+=M6}jTe_gq82o "2#>E=c&e' %,jM,́LDK( dD88 #t3DG H+tr | :A:1 H!m`aׁ-,_DS.:DחϜ1<9Z04[oNKzء)3- USftCMaȖƩlq:̅fQt@նd5l٥Y[qY`sVoBWo#, )k:=ը+ r,+g?'l@im?pyw\m\4:1XEJ"*<Ծֹ ̤(8HP^ֹ_ȳў[m_̰<qX l31c̓LifnXQi#pSзÈ4U..3զaY}W˳t⊄gW̿~>ExRjÐ3_j:#ِS()96,%QWDP.,љU"m}!I)ıCMYO3:~'FC Z krx^ﮯaj!oo*c_'fklvۃ_GRZ(a2Գޢtv=1vFceIŁjW>ZHv]؋.No=ݮ Kl}k$_[5&TI"r+m*UBi}HABXgaGWp>Gr l3% TOG!3LG A6鏑T_2v8NX*dA筜Fġ]Vݦ!@*XY~뭷]ѤoHZK B{ۖmz .:=FR#mF0*0SցJקN}EĦ@$+\mXW6)u$FLN:ZՌ{%}]qӹ5ǺDw'Ԭ>Cƣ4&nu>]\tDM&i3-#}OPvsHeOЀQGP ٔ"9m&N$f0覨egNKΏ+yÁKPudjbtn4I#I9B99mS&Ijo%Gv82 }ܘzlubNf5'1 V0#G ҬkmnAi/8t罾՜!$QP62TY2%Yk%}hxz:EFaӬg`.˓Dd]`)E<7kY%XjKMngVϡ%~9O`Dq+.b0$4uyiWY˖[yP/.9tb]1넑P'0!cr[j q1>ft`jCSpLkL.~͒χ~D<+W+w>~Z#oy8D=< Q|LX"o`;Sy'_vcN y(S$| sI\q1KNg)2%CD_c'\-WwOtq8P~!<v3irxH?6_ ͑-Os1;I"#wh4$)L>3ڵc; fC>f$$0EHdm$)>K(!@BsdJr}B_sPёP$ 2LXDྸT2 Ң 4(Ms*搋5>Or'b~q=D[ÓAlJԵ3DxLX2܎9p , gY=A_쳓+#.ώˎȷ5nۂ*q.DN@Q>㘡V&Ij6J7x0m$6)B+fv9(IU3Zt!Nd#IxVC1 y #]'Gjtݹt\ΐK0$ft̓f ao|>VN>ݲnU+#zE;iZڈ) "!1[CbNbn Wv0>&k)Z@Ld9N,ն@[Ԃ-OقcZJ9W:B4gQ3S^^]_DLQb&cJqi8 %sIB籵iORT9si|9 /Y*Uʋ[d^b`?pb&t_H1:$Y_7kjKߏ ٞF44%68NoRX0`X2, Xrhj湍lJ /P`tdX2Y/E[o"Ww3[t$.|,K]Э2SΕWÅW]iFS:}:j P1X@Iloɬ`MAX/Ԕu1 z8Jަ= YIͯDhз`]-n~6+RtL:5!*Ujo3ck\m/%HiӤ́S i2pS}'DA:@#gLx\-C#vkdۀɑ0alAqXW*A VgWAqGX.'_՝&ZoE5r5[fH:'))1xo !5jtcNzG|.VwF_?'ҟή_"^R HiׅfXxw .?\ݑԤYVxI^3KG۳tUrϫG*JF:ko$DVi_-3%rfppVxZ16ڋBԿ*[0k[hLJ\(TrYE+u EdRdA,aWE=yH 4jaHj,ީT]כNY+6>*''"̇DyZ%\kދP8}z"mxT&M6(NiNFt9FzdB|-^MCnшRIy D$KF<ҙ􇹻 ZI00X8Yy7TO, "HE(eQ$}3 wcy+2VRjI+Tdd}\4 s(&{º쨩}4nؽh/W>U*cK> :D5F6왶G:j JaA;i`-Ե>4CC&45=ij>pT'179dt@i6wHwqH}L|*rE׬lIVF~őbf$Mwgr[&z'YU,:|ZxfbOh{INΘmjB+*`,2(H#M>XD5r 2T D:H!bg1Z$@._mu {M/8bҚu;F٩Edb Z X[ޱ,n(@$Tkr3 %j<֋,c|FDZ|iq3 5jqOd2DnDn*ɩznTwN+zG @.3R!;̭$y &OKc6'jl|ۄt*F5j%4~vMr* :Q8~/k?=_Wg\|VB0zADk"ۤtsKV=LC1u<2 ĆWoo?k$Ëwfh=ڡ$_W4c7Dr/tYBͬE?'>УA6L),O`^Ƚ:|YK(=Xr6V, ys75O4#I8 Ղ2od~tI %*ɱ76S>zƤ=l \ӾbхVX YM]Nm*wf#wa@-!8+]Ik F7zUK*g 2^|%=86, >T)\tKh0':6*R*;:ZpA1mR}e"W^ޜ^{;?eUՇvx>cREN9 twѪz*GK♢Q&9FT)E"SMveBYF{fӴ<1T~8gWJw @-a-0nkCZQ".3(+RbllRds1U~1OZDhB ߰z!Î@إ(e&K=#r;t/nq(8Y裸 "vGLM:9 [+D~xw~o(PZ^{Y{GMi@Pr&'sSpxɍ,<S1ګdQg5묽LwզV\;g I %vq-/]Qxq݋p@7vǨ`X%FI^2`C !h*mD@ZQP"Zvf 2 ؠBV%RjFumv; w)% .&c' vc2\V@>{{aھxB`"FtJe;]ڟMˊdf[#:x10Mm/Vqk(94 3g8Q@z1qh7@!ife0R}6*;Z`[2A ٔCPC Zw/>KծQ[2 "U<*/oddN@* *+bLrbN"6bQd.1f:SMX+`OGP-VƗiLuLurj!GzbE1f+p$"TQ[eĢ)hcHa.TQX#ӡo#=d-9> .`慖X\ʫAKICLZ@%*Mnr6lf>PԖ곸9l9wY2ʑ$.Ԝ@=@mU狐Ujb-/ 2}zass_nL}3}ߍ2*ă'A0mqZ_5c0{fr@5%THJ1vlUZ_YkPb՛=k qX;\AVbyz{vSeWM"lj~j\IWM6DP>!:J)d!@x_P'\2<կI pǨtgV{@w>ece0&e`] lڳ/̉/`1,>F.<9Ўhv5j iw?E[N@Bz0]̮+9,{ힼO'rӋ\VICdQq>X~U$N.[w?ܔYkyiټOĖm#F"ďsxudx߻0 j 1Yӷ{GPw&lHML]9Eu2>AWrfyՀY8^r5%^uꖹKK26gc jYfN9.<2/*Fw>?{ RQ巫ׯ.ʃR8z Mz61&|tA3w={ܖ՟pW/=Ҧ:Ed3=ɑrW6s?iAfI/J!ƴ4;iƴlL;:dm_mC,g GgcHL^WdAyMFJUlupY]| >bGW{uvqN# ƁV~َ欵EJvIh#,!/1,c$Yj%кNꢔ7OvjU%Ch`Ȥ3~⁋ڴNئVjSyC1c6=-Gr5Ckɤ?&p9Ve]KA>$@._m#T!Ao+,! -z*P1J ^d*6)19\6)m,C*&Z_ ;4&jF?-;f}gcdE ^ +S\S",N]tFo:XǣYhϟGSd률+JAصdF8Gp*䋊0͗e%?!f e;|B6$k8D/養 *DDC19NN ˆ)&L%yqW˙ y L[z$ipnIں5z]w|GtD׽AlL:чI[//Gc)˸#mEX\S\Lf.JaJ!hqf%ճ O=s|6;zh$[U{}C=MGRJ}i}cP~8~?.p,F[w$A\0h%ĥ̺:CŒ%($DS>{d^t$"1# vXub)roBy,Wa. Gkr1m-#ŖHc$VboS`?j}QIݵ-X )a ʩl9&h܃aͨ6G%b<1g[ӡ#֏1`FaZݶw% &#(+. Zo.~a\Pϟcx|rvЍMkTÆ3+aUn]mC0AkS=lڃoR@7zT}߹3λ1䎹&EeS0d|RWWUW'ci$%VvzIsw["wY1/v$=H0 !3 .+TD\# ($YT$5:ANq,~d N&l+}?M(IJZ;:6? &[It3ADL_B)O2 pe:v{߿yͻdu?Zm,!}yBw&9x[;h6` __< ӋM[00<&+HagYʊd3!ߊ<[ȥtŴ*N"p>tlIR"R#]/Q)$X&lm' js#٥'e39b'IY-D2hT$~@Y$~Ym,):Υ_JN:J zitEqC( M܉<[#Xz!=xN(<!"Iꤋ-h'\#|G)^3ؿ4:Fu6as(+3P ͼ&YJ3`HH)"vCIКh I:qLԏ% XHm F{W&R$d 2zR9rEl\y0(SJ sL`[=P6OE6e2ݓ"J%Ng}N( )L X" 3T᰼e@J gAʁn4+R[WXC 2mHH%*, %h3ky&Dŀ@òLNe{*xОs4 fƩfR#q'>Z'`q!>A#ZKjiUTpw]]#L]!rSs1Kos K$'{0 g@~h.zSI3m p J=4>8ȅ[?qgKidy7Xf@b @GqB>AMk#ȹ@Vorz|X8/u jt Rغi0~?`b^-씓6zqܧz)jx'^t;Gj܂\- G&%A 0fSձY.e)BL4$ݶ̐#kt%iCQJ^|::YYI9Xp+tܰz}zW! uz8+"x ڳB1O[`nOobi̅ -~;PrN[Q[&9l[-zLj!Z@ǃp44>ZQؿ߂]'yD ? ﹡x؛}Yn&(HNX"7 %]M+Gȶx:lhQ/ѡ?Ցq> sj"OrY혗UQM%DSU7{I/ R}͹ɹrnDInNԭC0 gUmp5.WJKV&ax,VB+:Eg9 v \ iM5Z.mbQ4Ck%Vӭj#ERpD=g[ k7?lDBO&9f4>;_L(Sބd)-l8F" RӢk^2d ^-NJa5T!!p !sXгl2MsցB\"$a ` 4b_q2UpU\"'wSXMTݗlF5LnGH1';:pɣ)NaIIZ16ۈVYlaCy 3?m;oHY@&x%{dQAwPĐYK^ 'A!kL<[|FVVl6͸&Bx*ʷ+b姉 bξ(:pL,X0#N$< XqE\Y!AXDZW- u;7^$鲫Ln|-no]}kA…|ǽfjă3;clW(|GԆ\pE u&p< Gr1q`:eȻoՃvQ_![q< u|3thnʈz3Vjm_vMwv$9F2RzBYxmwo QTvcNI"u\k<:hષN{)u0mLg=8tޞCQ\ho`JL( 4'3%VA(dZ:)2p"Gl>ym3& \`IT!`Ҥ"gv t ų W\KۛVueoŗbѼG}b*%4j+nKorV琾 S*},:*/UyA' PP/AY`$'`g|G+ѓUn,0bG;? Kq{;n"tPE :V#0Fδk՚M[jAz!-#\} ?~8yG9û5ft8kq|͍>JLk^<AH ~=fSYxu>klS(2[cqus>L\|957cD4Ui~ߟD4̽ojЇѽ_͍[;ǃ8DAeI`b3imf(#fޱLCBxf ],  z:v+ ŏJ,.i?Btşb<ȣ|ˡ9:mdٿrMvϱ̼a8$[(xclb<=0dYXh=ӏc,k+[4g$޵62 *ή#$bPQզ J>8R&ߪg(vLQu1g{Mw9&SO\MitFD(¨JPq0fl$H40 4Ξ@ VmWI]b5,D,2QQ$PA 5!D@A)u2jP&A-jh„-frB7G5M˛Y8%Ek/>ϗ 1&4hr`P%`)wIw:rgVPGGS[le Ce?wh%Q:.8y4h&ӷӝrQP C0A:~V\u(ED.t;hb$E7yv;e])H)(#РLq/:b5Eȋ'ubX>[>ʴkjT:,%` _C?JG7ou!ɆnȔHh$aWB՚-Y6 :9–֕T4,?G%ǥz>ε귽Ϧ#փuZ~vnv3-V}dEJ{ l&LwGP($g80FQTA;Zڳd㤩Eko1ER˻:V&ҥS%;H%`p+K4$CQ=*Y TCbj6HlH8 ֌DI,$7)8u u_[}T`*vԆ]yoetu6vo1&rEw I]*lͦ6׌lPdM fL ¡Ѓm/;;gF\J]nź2Xn9~oo'V;>$I %OQ:; @+,gd7Vfх~z׺[y1lMD;NJsV`|W^$ժNfE9e_/yoVE A4fN 0y[$d}:F%4dܦ(b ! J)ER DP)$I+8&8$:4M(cI(L$TUdi2\;!hݠxuxV\|0x`h(4(XHydi&6ÉVɘ( xA G@h}52YD"IMHqID`WAD S(O` {2$D&)7$TDnG-vcB`?fyc'ݚ;./kRN:㤩fRj{aY>jZcGk;X~0%ƒlY)aNJF_Zkߒ R;w8 1 :9c戭+ťk]M`H5М2!32 qLB)5ՠӌ3rB 6$ 5"Ĉ_ uZevow5ǏcIY@`0zi:  ![QMjH,gw\]r>d5зU5QBћ2lxYvk>W,"n%jmM2ٺ7i>"rw^ݽY8 AǶtO2ogfLmN"wv502l2lc.kEi>ו+~Ujq«sy낈nVႈN] `4R@E%Tl 1g7>-qU _6s3*bAax=G׊GF3kCBkeqWfn]I)x?]'Lt,3@t_~TPy|닝/6ݸYÆXYrHxev6w@7N-U@%F v# `aJFAcC^w~ Gg'o;^+6*B~Bg ? MB3i`[ŧ:rac0P"&PrPCdAtɐɐ ~21|Ɔ%MPwBG% u!6C<۶_ n]/ 9JF Q .%#݌d$#u V[KHd̠DG%;j|VR!,2JNTꕆ#LDH`H RÐ"19C$xQEKl:;-;"pQZs@ӱvƠ?ErnDyT.x1zr58>Zl-e[FHY =oRQ-Hjw; 4-%=VnZ6a\ lY!>i ג|۳W²ˠ^,y(úc q񩖪BYAL7yDB<J"JEgɅIWtim bsD*[-n9=ȍB>p#m,P F~yiK+pGTuW~LsFјʤtJw?Z(ˉgW6|Lbyި"kK:shc}X_]I&b4Nزec˖e,zFxQp%RhƉ-X$&1fav0Apas`]n{Fij~hx 1и&2׀oz@WfAP st9\. ӹ-)`i)MPw硆ɣ>9HɅDMtt:~WZU UXRE#Xy !%tkub[i&Zr=oY"&)fBLAϙ(dc~'*-]:^H^'hJ(A"P,(ɲ'C >2Ht76R5eWr]U]V%JjrSu3U`߉H>΋j`f(w ԱmgCIC-(~ N{ȦXf bd?)C-}q$qMumpذ8?q/oOGWOE&Q3oH{%ۭ7 Ay~/xg 8/[5{'{,G[J{nP5߻LJ^GzWHJz&:zq/gcY{Ku /~0cX+tt ""ٟioRhξ, cxkt-G7;;Ļ8c~ 0[FUSAȞFybKg/_esd_i.o+ހ^cXU_6Vz?|=9x-'nŪev}s=xVGW/w;~͑iKci;=OOx>WG{{P h3*nw^~A^.VD&3|ZD}c3{Ao>o߷>nۯA7؛W{|/ɼ:flg)~rI.w'x3IdKC.v_?ϸ=Ͷ3ul3eɝ?z_ͷɽ<ޫt㲝 kۖ}:&nO~1abhӋv3}Z&G\zʓٍ |lEwEwb ߶T;k᣼Shhx|L9ݯgo qgss`:OҾ M97:yҭl} o4eܴy 47[/&Xs8x @,xڲrS5{ѷɫ!jRc`3o u=t~ "S\o B6'orpzAhb!=d6`y)g[פ˒<*XJOU{ x'h>-4ČtB9^7!rl'cn$ FM&:b-jcjƁ:m}MBfhmiUxw@yV Thlg |mp60B8WV(_@|<єdYr!"ѥ-RSdQ6nL $ڪk{RsT>c~oM/Kk}?.bbʏ{[sCQWpf=|w^T|?]g7Hoxw5_<koV]6翾lzu)Y<|Ps""+H3[n.ie-o1i宥UU:bR5|tM2 &l =04Q{xa'ʍKtyH>=G00-^$%a4Kx>qᏴfҧ(O߫e3 txpWMIZɂ_Eǘ-4, ʊZ}-qdE]dګBGaêfij7t֟} _0vl(:o]t8W"Ȟqs>^7`:OǤNJm M_ߖn-gwQK ڨYC\P*"|>l顽׺mݚ N/Z׻'!\C-]_F= ߐ`d1]o"+BΨ]R3c ;ӐYK9l7֩S/q6 m6Ah3Rٳj22s,l-t)^Ry9z jI.>8^|MUoUP-[rI9p)v%%ޣ~pW(c8A=\E{bPgh:ĠaZ2ʓϬ ۙ3-Y*:y (BlTZ%&w/;j;g*aE`t8tYb$u9F6[#ǃuϝ+Qx4  iAbeYR4f.Ӎ :]I#jbD_6=l'c_7!2GS\5vnm1Qs~K*jw{U/uV`αw;˿-Sׁ(Q!ӅrF;c % pVmh lձB*e:Ь(s5q2B]uH7F0[;@Vn)J@礭xN:T{ҏm*w)a9PvƠ6qv:8MKP?ӁLs8e{—Vy@pjamYΫ1?ᖣW0Ӹ  M[1T2\DaL5{0ԓèc2Ǹ%_‰3xYʗJ?? i|%Y4E-NqÅ_ӫ͌Kڊjv `{l stn ey̜;(yF '܊~!3숖gbsѲN+e'--*;KGLwWMXO>j=mi "G{ߔJ>J癳EZ% 2Uڹٜ'iIj8iP남htʽdR˲l{T83+wZJa?˻^n:-OUj rғҰr~REXK´Zk>kik)0-*#k)UVTO[KUiPQh <+-[8kIk)0-%lrLVTs?InLC%rJ硸t~Q,;uo|?aN0kkx@5@ Z0Tr 6gŶLH"BLTAJ%]zSJM,Z(1 XXZH0rRڹ+I3ȕ >BLf~liio3h=8=+[Â6M\/8cjf :/ùQTW@N5$c̫IWmTIRmo~ Z-h3n3E`,p_=MK -%#6var/home/core/zuul-output/logs/kubelet.log0000644000000000000000005603310515156305154017705 0ustar rootrootMar 17 15:36:47 crc systemd[1]: Starting Kubernetes Kubelet... Mar 17 15:36:47 crc restorecon[4668]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:47 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:48 crc restorecon[4668]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 15:36:48 crc restorecon[4668]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 17 15:36:49 crc kubenswrapper[4767]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 15:36:49 crc kubenswrapper[4767]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 17 15:36:49 crc kubenswrapper[4767]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 15:36:49 crc kubenswrapper[4767]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 15:36:49 crc kubenswrapper[4767]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 17 15:36:49 crc kubenswrapper[4767]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.108932 4767 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114140 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114210 4767 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114223 4767 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114234 4767 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114262 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114276 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114285 4767 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114297 4767 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114307 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114317 4767 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114327 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114337 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114346 4767 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114356 4767 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114367 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114376 4767 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114385 4767 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114394 4767 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114403 4767 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114413 4767 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114423 4767 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114433 4767 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114442 4767 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114450 4767 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114461 4767 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114474 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114486 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114499 4767 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114510 4767 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114520 4767 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114532 4767 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114543 4767 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114552 4767 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114562 4767 feature_gate.go:330] unrecognized feature gate: Example Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114571 4767 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114581 4767 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114589 4767 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114597 4767 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114606 4767 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114615 4767 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114622 4767 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114630 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114638 4767 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114645 4767 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114653 4767 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114660 4767 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114669 4767 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114677 4767 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114685 4767 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114692 4767 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114700 4767 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114710 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114718 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114726 4767 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114735 4767 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114743 4767 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114751 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114759 4767 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114767 4767 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114774 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114782 4767 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114789 4767 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114798 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114806 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114816 4767 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114824 4767 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114832 4767 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114840 4767 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114848 4767 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114855 4767 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.114863 4767 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115079 4767 flags.go:64] FLAG: --address="0.0.0.0" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115097 4767 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115110 4767 flags.go:64] FLAG: --anonymous-auth="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115121 4767 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115153 4767 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115162 4767 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115196 4767 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115207 4767 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115217 4767 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115226 4767 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115236 4767 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115245 4767 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115254 4767 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115265 4767 flags.go:64] FLAG: --cgroup-root="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115274 4767 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115283 4767 flags.go:64] FLAG: --client-ca-file="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115291 4767 flags.go:64] FLAG: --cloud-config="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115300 4767 flags.go:64] FLAG: --cloud-provider="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115309 4767 flags.go:64] FLAG: --cluster-dns="[]" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115321 4767 flags.go:64] FLAG: --cluster-domain="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115329 4767 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115338 4767 flags.go:64] FLAG: --config-dir="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115347 4767 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115357 4767 flags.go:64] FLAG: --container-log-max-files="5" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115367 4767 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115376 4767 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115385 4767 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115395 4767 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115404 4767 flags.go:64] FLAG: --contention-profiling="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115413 4767 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115422 4767 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115431 4767 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115440 4767 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115451 4767 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115462 4767 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115471 4767 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115481 4767 flags.go:64] FLAG: --enable-load-reader="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115490 4767 flags.go:64] FLAG: --enable-server="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115500 4767 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115511 4767 flags.go:64] FLAG: --event-burst="100" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115521 4767 flags.go:64] FLAG: --event-qps="50" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115530 4767 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115539 4767 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115548 4767 flags.go:64] FLAG: --eviction-hard="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115558 4767 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115567 4767 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115576 4767 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115585 4767 flags.go:64] FLAG: --eviction-soft="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115594 4767 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115603 4767 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115612 4767 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115621 4767 flags.go:64] FLAG: --experimental-mounter-path="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115629 4767 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115638 4767 flags.go:64] FLAG: --fail-swap-on="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115647 4767 flags.go:64] FLAG: --feature-gates="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115657 4767 flags.go:64] FLAG: --file-check-frequency="20s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115667 4767 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115676 4767 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115686 4767 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115696 4767 flags.go:64] FLAG: --healthz-port="10248" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115704 4767 flags.go:64] FLAG: --help="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115713 4767 flags.go:64] FLAG: --hostname-override="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115722 4767 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115732 4767 flags.go:64] FLAG: --http-check-frequency="20s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115741 4767 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115750 4767 flags.go:64] FLAG: --image-credential-provider-config="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115758 4767 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115767 4767 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115775 4767 flags.go:64] FLAG: --image-service-endpoint="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115784 4767 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115793 4767 flags.go:64] FLAG: --kube-api-burst="100" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115801 4767 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115811 4767 flags.go:64] FLAG: --kube-api-qps="50" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115819 4767 flags.go:64] FLAG: --kube-reserved="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115828 4767 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115837 4767 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115846 4767 flags.go:64] FLAG: --kubelet-cgroups="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115854 4767 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115863 4767 flags.go:64] FLAG: --lock-file="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115872 4767 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115881 4767 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115890 4767 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115904 4767 flags.go:64] FLAG: --log-json-split-stream="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115913 4767 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115922 4767 flags.go:64] FLAG: --log-text-split-stream="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115931 4767 flags.go:64] FLAG: --logging-format="text" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115940 4767 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115949 4767 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115958 4767 flags.go:64] FLAG: --manifest-url="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115966 4767 flags.go:64] FLAG: --manifest-url-header="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115978 4767 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115987 4767 flags.go:64] FLAG: --max-open-files="1000000" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.115997 4767 flags.go:64] FLAG: --max-pods="110" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116006 4767 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116015 4767 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116025 4767 flags.go:64] FLAG: --memory-manager-policy="None" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116034 4767 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116043 4767 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116052 4767 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116062 4767 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116080 4767 flags.go:64] FLAG: --node-status-max-images="50" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116089 4767 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116098 4767 flags.go:64] FLAG: --oom-score-adj="-999" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116107 4767 flags.go:64] FLAG: --pod-cidr="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116115 4767 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116129 4767 flags.go:64] FLAG: --pod-manifest-path="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116138 4767 flags.go:64] FLAG: --pod-max-pids="-1" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116147 4767 flags.go:64] FLAG: --pods-per-core="0" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116156 4767 flags.go:64] FLAG: --port="10250" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116164 4767 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116195 4767 flags.go:64] FLAG: --provider-id="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116204 4767 flags.go:64] FLAG: --qos-reserved="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116213 4767 flags.go:64] FLAG: --read-only-port="10255" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116222 4767 flags.go:64] FLAG: --register-node="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116230 4767 flags.go:64] FLAG: --register-schedulable="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116239 4767 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116254 4767 flags.go:64] FLAG: --registry-burst="10" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116262 4767 flags.go:64] FLAG: --registry-qps="5" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116271 4767 flags.go:64] FLAG: --reserved-cpus="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116279 4767 flags.go:64] FLAG: --reserved-memory="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116291 4767 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116300 4767 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116308 4767 flags.go:64] FLAG: --rotate-certificates="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116317 4767 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116326 4767 flags.go:64] FLAG: --runonce="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116335 4767 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116344 4767 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116354 4767 flags.go:64] FLAG: --seccomp-default="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116362 4767 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116371 4767 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116380 4767 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116389 4767 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116398 4767 flags.go:64] FLAG: --storage-driver-password="root" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116407 4767 flags.go:64] FLAG: --storage-driver-secure="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116416 4767 flags.go:64] FLAG: --storage-driver-table="stats" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116426 4767 flags.go:64] FLAG: --storage-driver-user="root" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116436 4767 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116446 4767 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116455 4767 flags.go:64] FLAG: --system-cgroups="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116465 4767 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116478 4767 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116487 4767 flags.go:64] FLAG: --tls-cert-file="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116496 4767 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116506 4767 flags.go:64] FLAG: --tls-min-version="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116515 4767 flags.go:64] FLAG: --tls-private-key-file="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116524 4767 flags.go:64] FLAG: --topology-manager-policy="none" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116532 4767 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116542 4767 flags.go:64] FLAG: --topology-manager-scope="container" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116551 4767 flags.go:64] FLAG: --v="2" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116562 4767 flags.go:64] FLAG: --version="false" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116574 4767 flags.go:64] FLAG: --vmodule="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116584 4767 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.116594 4767 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116800 4767 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116810 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116820 4767 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116830 4767 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116840 4767 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116849 4767 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116858 4767 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116866 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116875 4767 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116883 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116891 4767 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116898 4767 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116906 4767 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116914 4767 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116922 4767 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116929 4767 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116937 4767 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116945 4767 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116953 4767 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116961 4767 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116969 4767 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116977 4767 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116985 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.116992 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117000 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117009 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117017 4767 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117024 4767 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117034 4767 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117045 4767 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117067 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117096 4767 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117105 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117113 4767 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117122 4767 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117132 4767 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117141 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117150 4767 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117157 4767 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117165 4767 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117201 4767 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117210 4767 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117218 4767 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117226 4767 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117233 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117243 4767 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117254 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117262 4767 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117270 4767 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117278 4767 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117286 4767 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117294 4767 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117302 4767 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117309 4767 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117318 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117326 4767 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117333 4767 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117341 4767 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117350 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117358 4767 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117366 4767 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117374 4767 feature_gate.go:330] unrecognized feature gate: Example Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117385 4767 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117392 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117400 4767 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117407 4767 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117415 4767 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117423 4767 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117430 4767 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117438 4767 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.117445 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.118354 4767 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.131963 4767 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.132003 4767 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132122 4767 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132133 4767 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132140 4767 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132147 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132155 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132162 4767 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132195 4767 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132204 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132210 4767 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132217 4767 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132225 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132231 4767 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132237 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132243 4767 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132250 4767 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132256 4767 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132262 4767 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132269 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132275 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132281 4767 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132287 4767 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132296 4767 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132306 4767 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132312 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132319 4767 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132325 4767 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132334 4767 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132343 4767 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132350 4767 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132356 4767 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132363 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132369 4767 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132377 4767 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132383 4767 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132389 4767 feature_gate.go:330] unrecognized feature gate: Example Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132396 4767 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132403 4767 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132411 4767 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132419 4767 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132427 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132434 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132442 4767 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132449 4767 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132456 4767 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132463 4767 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132469 4767 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132476 4767 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132482 4767 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132489 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132495 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132501 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132508 4767 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132515 4767 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132521 4767 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132527 4767 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132536 4767 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132544 4767 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132551 4767 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132558 4767 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132567 4767 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132573 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132579 4767 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132586 4767 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132595 4767 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132603 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132611 4767 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132618 4767 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132624 4767 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132632 4767 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132640 4767 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132648 4767 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.132660 4767 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132864 4767 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132877 4767 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132883 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132891 4767 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132898 4767 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132907 4767 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132917 4767 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132924 4767 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132932 4767 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132940 4767 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132951 4767 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132962 4767 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132970 4767 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132977 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132985 4767 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.132993 4767 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133000 4767 feature_gate.go:330] unrecognized feature gate: Example Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133008 4767 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133015 4767 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133021 4767 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133028 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133033 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133038 4767 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133045 4767 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133054 4767 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133063 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133070 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133077 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133116 4767 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133123 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133130 4767 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133138 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133145 4767 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133151 4767 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133157 4767 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133164 4767 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133198 4767 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133205 4767 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133212 4767 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133218 4767 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133224 4767 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133230 4767 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133237 4767 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133243 4767 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133249 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133256 4767 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133262 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133269 4767 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133275 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133282 4767 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133288 4767 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133294 4767 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133300 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133306 4767 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133312 4767 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133319 4767 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133325 4767 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133331 4767 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133338 4767 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133344 4767 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133351 4767 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133357 4767 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133363 4767 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133370 4767 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133377 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133383 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133389 4767 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133395 4767 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133401 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133408 4767 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.133414 4767 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.133424 4767 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.134527 4767 server.go:940] "Client rotation is on, will bootstrap in background" Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.141440 4767 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.147255 4767 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.147477 4767 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.149613 4767 server.go:997] "Starting client certificate rotation" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.149654 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.149883 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.181509 4767 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.184835 4767 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.185333 4767 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.204892 4767 log.go:25] "Validated CRI v1 runtime API" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.248272 4767 log.go:25] "Validated CRI v1 image API" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.250465 4767 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.256230 4767 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-17-15-18-52-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.256276 4767 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.274743 4767 manager.go:217] Machine: {Timestamp:2026-03-17 15:36:49.272848582 +0000 UTC m=+0.686164649 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8b268e8f-27e3-49d0-9d37-42af1be2e9f2 BootID:6bc6e9eb-1e1c-467d-b62d-b25deb35366f Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c7:74:ed Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:c7:74:ed Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:62:30:58 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:47:f6:c9 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:66:56:1b Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:26:6c:57 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:02:00:9a:9c:7d:a4 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:9e:dd:0e:f1:f4:77 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.274988 4767 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.275117 4767 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.275834 4767 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.276060 4767 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.276095 4767 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.276485 4767 topology_manager.go:138] "Creating topology manager with none policy" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.276495 4767 container_manager_linux.go:303] "Creating device plugin manager" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.276964 4767 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.276995 4767 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.277131 4767 state_mem.go:36] "Initialized new in-memory state store" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.277707 4767 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.283792 4767 kubelet.go:418] "Attempting to sync node with API server" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.283822 4767 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.283852 4767 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.283872 4767 kubelet.go:324] "Adding apiserver pod source" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.283896 4767 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.288503 4767 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.289801 4767 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.292912 4767 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.293305 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.293391 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.293453 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.293531 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294460 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294498 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294510 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294523 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294541 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294552 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294563 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294580 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294593 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294609 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294626 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.294637 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.295747 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.296260 4767 server.go:1280] "Started kubelet" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.296458 4767 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.297477 4767 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 17 15:36:49 crc systemd[1]: Started Kubernetes Kubelet. Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.303634 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.304477 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.304543 4767 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.304570 4767 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.305450 4767 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.305497 4767 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.305600 4767 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.306755 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.306926 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" interval="200ms" Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.307880 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.308307 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.309731 4767 factory.go:55] Registering systemd factory Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.309827 4767 factory.go:221] Registration of the systemd container factory successfully Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.310227 4767 factory.go:153] Registering CRI-O factory Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.310315 4767 factory.go:221] Registration of the crio container factory successfully Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.310445 4767 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.310528 4767 factory.go:103] Registering Raw factory Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.310618 4767 manager.go:1196] Started watching for new ooms in manager Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.311574 4767 server.go:460] "Adding debug handlers to kubelet server" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.312260 4767 manager.go:319] Starting recovery of all containers Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.316759 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.119:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189daaf7bf416d9e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,LastTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326295 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326390 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326421 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326448 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326487 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326515 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326541 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326567 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326597 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326627 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326652 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326677 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326703 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326747 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326774 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.326804 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329335 4767 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329404 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329433 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329459 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329487 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329512 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329540 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329568 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329593 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329619 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329646 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329681 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329708 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329736 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329762 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329789 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329870 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329899 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329923 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.329982 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330012 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330036 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330059 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330083 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330110 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330137 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330165 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330491 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330517 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330543 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330570 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330596 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330626 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330650 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330677 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330703 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330729 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330764 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330808 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330838 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330866 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330897 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330924 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.330958 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331008 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331034 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331062 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331090 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331117 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331142 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331167 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331230 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331257 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331283 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331308 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331336 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331365 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331391 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331418 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331455 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331484 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331509 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331539 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331565 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331593 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331621 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331647 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331673 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331770 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331814 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331844 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331871 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331903 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331930 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331958 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.331985 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332012 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332039 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332065 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332094 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332121 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332150 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332212 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332240 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332268 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332295 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332323 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332349 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332377 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332415 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332444 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332476 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332508 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332537 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332564 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332594 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332626 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332657 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332683 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332709 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332736 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332761 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332790 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332817 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332842 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332871 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332896 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332920 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332948 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.332976 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333002 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333032 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333062 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333089 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333114 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333142 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333230 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333311 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333343 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333373 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333402 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333428 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333454 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333483 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333510 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333536 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333563 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333588 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333619 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333647 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333673 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333699 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333725 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333752 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333777 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333804 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333829 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333853 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333880 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333906 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333932 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333960 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.333983 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334009 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334037 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334062 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334087 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334121 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334145 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334164 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334223 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334247 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334474 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334495 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334514 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334532 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334550 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334569 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334585 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334604 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334624 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334641 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334658 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334675 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334691 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334709 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334727 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334746 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334764 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334780 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334798 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334815 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334832 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334849 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334869 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334888 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334908 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334926 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334943 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334961 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334978 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.334997 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335014 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335031 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335049 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335065 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335083 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335100 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335117 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335136 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335156 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335206 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335232 4767 reconstruct.go:97] "Volume reconstruction finished" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.335245 4767 reconciler.go:26] "Reconciler: start to sync state" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.338752 4767 manager.go:324] Recovery completed Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.349815 4767 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.351012 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.352162 4767 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.352997 4767 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.353043 4767 kubelet.go:2335] "Starting kubelet main sync loop" Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.353120 4767 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.353540 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.353772 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.353803 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.355484 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.355630 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.356664 4767 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.356756 4767 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.356839 4767 state_mem.go:36] "Initialized new in-memory state store" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.374683 4767 policy_none.go:49] "None policy: Start" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.376523 4767 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.376554 4767 state_mem.go:35] "Initializing new in-memory state store" Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.407059 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.444744 4767 manager.go:334] "Starting Device Plugin manager" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.445396 4767 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.445426 4767 server.go:79] "Starting device plugin registration server" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.445785 4767 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.445801 4767 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.445972 4767 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.446043 4767 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.446051 4767 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.453611 4767 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.453797 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.455239 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.455281 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.455295 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.455461 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.455913 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.455995 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.456305 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.456350 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.456362 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.456537 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.456713 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.456770 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.457428 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.457464 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.457489 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.457440 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.457578 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.457591 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.457751 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.457777 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.457794 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.458290 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.458830 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.458854 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.458865 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.458871 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.458888 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.458921 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.458938 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.458894 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.459045 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.459278 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.459374 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.459413 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.460378 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.460409 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.460429 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.460440 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.460412 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.460548 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.460561 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.460588 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.461619 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.461650 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.461661 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.507876 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" interval="400ms" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537288 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537355 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537377 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537396 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537413 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537430 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537448 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537466 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537484 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537499 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537515 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537531 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537550 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537566 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.537587 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.547143 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.548221 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.548254 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.548262 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.548282 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.549448 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.119:6443: connect: connection refused" node="crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.638924 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639066 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639118 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639133 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639232 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639263 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639328 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639353 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639397 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639418 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639454 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639589 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639662 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639557 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639503 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639595 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639645 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639587 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639774 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639808 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639883 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639906 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639929 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639950 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639971 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.639994 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.640020 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.640040 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.640061 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.640546 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.750598 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.752331 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.752399 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.752420 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.752479 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.753378 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.119:6443: connect: connection refused" node="crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.802163 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.819751 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.833997 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.858498 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.861019 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-0052b67c52284441569748edc6c40e3bf431161d06a57f50f3caee037d164f95 WatchSource:0}: Error finding container 0052b67c52284441569748edc6c40e3bf431161d06a57f50f3caee037d164f95: Status 404 returned error can't find the container with id 0052b67c52284441569748edc6c40e3bf431161d06a57f50f3caee037d164f95 Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.861647 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-a18c108259ced91e2a9c7038caf620745e8eacad29e75c3922f098e620c7c5b8 WatchSource:0}: Error finding container a18c108259ced91e2a9c7038caf620745e8eacad29e75c3922f098e620c7c5b8: Status 404 returned error can't find the container with id a18c108259ced91e2a9c7038caf620745e8eacad29e75c3922f098e620c7c5b8 Mar 17 15:36:49 crc kubenswrapper[4767]: I0317 15:36:49.863396 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.866328 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-c3d2f7262798f4b93581877ba0521979553f99c0496a95ec9a8c1bdc09638a13 WatchSource:0}: Error finding container c3d2f7262798f4b93581877ba0521979553f99c0496a95ec9a8c1bdc09638a13: Status 404 returned error can't find the container with id c3d2f7262798f4b93581877ba0521979553f99c0496a95ec9a8c1bdc09638a13 Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.876244 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-86c0daaf310d5619895187f508a552d535e67c8fc638ea5551faee6230339411 WatchSource:0}: Error finding container 86c0daaf310d5619895187f508a552d535e67c8fc638ea5551faee6230339411: Status 404 returned error can't find the container with id 86c0daaf310d5619895187f508a552d535e67c8fc638ea5551faee6230339411 Mar 17 15:36:49 crc kubenswrapper[4767]: W0317 15:36:49.883561 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-29af384e6a710b6f289e366e0dd16cb3965cbaeaa6232c86fa11151751006d0f WatchSource:0}: Error finding container 29af384e6a710b6f289e366e0dd16cb3965cbaeaa6232c86fa11151751006d0f: Status 404 returned error can't find the container with id 29af384e6a710b6f289e366e0dd16cb3965cbaeaa6232c86fa11151751006d0f Mar 17 15:36:49 crc kubenswrapper[4767]: E0317 15:36:49.909379 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" interval="800ms" Mar 17 15:36:50 crc kubenswrapper[4767]: W0317 15:36:50.145077 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:50 crc kubenswrapper[4767]: E0317 15:36:50.145162 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.154003 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.155138 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.155192 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.155203 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.155227 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:36:50 crc kubenswrapper[4767]: E0317 15:36:50.155554 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.119:6443: connect: connection refused" node="crc" Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.306717 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.358450 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a18c108259ced91e2a9c7038caf620745e8eacad29e75c3922f098e620c7c5b8"} Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.359288 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0052b67c52284441569748edc6c40e3bf431161d06a57f50f3caee037d164f95"} Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.360114 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"29af384e6a710b6f289e366e0dd16cb3965cbaeaa6232c86fa11151751006d0f"} Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.360796 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"86c0daaf310d5619895187f508a552d535e67c8fc638ea5551faee6230339411"} Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.361539 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c3d2f7262798f4b93581877ba0521979553f99c0496a95ec9a8c1bdc09638a13"} Mar 17 15:36:50 crc kubenswrapper[4767]: E0317 15:36:50.469036 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.119:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189daaf7bf416d9e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,LastTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:36:50 crc kubenswrapper[4767]: E0317 15:36:50.710556 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" interval="1.6s" Mar 17 15:36:50 crc kubenswrapper[4767]: W0317 15:36:50.730163 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:50 crc kubenswrapper[4767]: E0317 15:36:50.730342 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:50 crc kubenswrapper[4767]: W0317 15:36:50.757733 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:50 crc kubenswrapper[4767]: E0317 15:36:50.757876 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:50 crc kubenswrapper[4767]: W0317 15:36:50.858520 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:50 crc kubenswrapper[4767]: E0317 15:36:50.858653 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.956120 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.958209 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.958252 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.958261 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:50 crc kubenswrapper[4767]: I0317 15:36:50.958285 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:36:50 crc kubenswrapper[4767]: E0317 15:36:50.958569 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.119:6443: connect: connection refused" node="crc" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.216505 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 15:36:51 crc kubenswrapper[4767]: E0317 15:36:51.217660 4767 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.307243 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.367365 4767 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46" exitCode=0 Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.367488 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.367475 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46"} Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.368637 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.368705 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.368723 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.369536 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f" exitCode=0 Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.369629 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f"} Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.369663 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.371040 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.371066 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.371074 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.373698 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"430edad2863573f2d2069c23335b51f5100526f65ab2c9ff799607ed1dbda8ed"} Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.373731 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0c13faf3ff976f76a211bdbb23b12d1165f4905f1a20b7afd81e9250f4b9c792"} Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.373770 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bc151899c4826bb8672c4ffbc2b1a199e2193f4f5e9fa86f6c68565ff7cf0c68"} Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.373775 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.373782 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1a6d14de833771c3a060a2879802f4f7f37b96b60004b4e5f96374e5fee1428e"} Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.374243 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.374604 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.374663 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.374683 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.375076 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.375106 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.375117 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.375704 4767 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e" exitCode=0 Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.375766 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e"} Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.375833 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.376543 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.376577 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.376615 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.377938 4767 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241" exitCode=0 Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.378015 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.378011 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241"} Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.378870 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.378922 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:51 crc kubenswrapper[4767]: I0317 15:36:51.378935 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.306854 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:52 crc kubenswrapper[4767]: E0317 15:36:52.311555 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" interval="3.2s" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.331211 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.379215 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.382634 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff"} Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.382673 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0"} Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.382684 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29"} Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.382693 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f"} Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.384339 4767 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2" exitCode=0 Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.384419 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2"} Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.384489 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.385262 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.385285 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.385294 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.386713 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2979605c7105c9f289cbfc223356236cea8fb72fec0b0242dea17fb1ad10e950"} Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.386745 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.387507 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.387538 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.387549 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.389195 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.389217 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.389192 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c"} Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.389302 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f"} Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.389317 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb"} Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.389330 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.390143 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.390183 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.390193 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.392562 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.392589 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.392597 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.559234 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.560308 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.560340 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.560354 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.560377 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:36:52 crc kubenswrapper[4767]: E0317 15:36:52.560782 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.119:6443: connect: connection refused" node="crc" Mar 17 15:36:52 crc kubenswrapper[4767]: W0317 15:36:52.591794 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:52 crc kubenswrapper[4767]: E0317 15:36:52.591890 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:52 crc kubenswrapper[4767]: I0317 15:36:52.783867 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:53 crc kubenswrapper[4767]: W0317 15:36:53.117359 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.119:6443: connect: connection refused Mar 17 15:36:53 crc kubenswrapper[4767]: E0317 15:36:53.117480 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.119:6443: connect: connection refused" logger="UnhandledError" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.393471 4767 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029" exitCode=0 Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.393561 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029"} Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.393603 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.394917 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.394963 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.394976 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.397544 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.400112 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="59588a28aef91c73a40b25c5a32c3fc6e7bdd42fcc66c9fb79efd9165c68db71" exitCode=255 Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.400257 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.400279 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.400328 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.400350 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.400281 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"59588a28aef91c73a40b25c5a32c3fc6e7bdd42fcc66c9fb79efd9165c68db71"} Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.400456 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402007 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402032 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402041 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402069 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402047 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402089 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402076 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402115 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402138 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402143 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402092 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.402148 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:53 crc kubenswrapper[4767]: I0317 15:36:53.403872 4767 scope.go:117] "RemoveContainer" containerID="59588a28aef91c73a40b25c5a32c3fc6e7bdd42fcc66c9fb79efd9165c68db71" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.407073 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.409930 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c214a1fd1674cbd43fd386473a8e997ca090b74b2e6b7cec66393dafdd058d6f"} Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.410097 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.410767 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.411204 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.411241 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.411258 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.414444 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3"} Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.414494 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd"} Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.414513 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12"} Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.414530 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429"} Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.414549 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.414716 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.415630 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.415685 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.415694 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.416119 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.416161 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:54 crc kubenswrapper[4767]: I0317 15:36:54.416199 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.322489 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.326130 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.422378 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514"} Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.422678 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.422746 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.423571 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.423612 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.424715 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.424753 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.424765 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.424718 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.424856 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.424898 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.426111 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.426152 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.426242 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.761641 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.762846 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.762882 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.762897 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:55 crc kubenswrapper[4767]: I0317 15:36:55.762919 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:36:56 crc kubenswrapper[4767]: I0317 15:36:56.425094 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:56 crc kubenswrapper[4767]: I0317 15:36:56.425364 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:56 crc kubenswrapper[4767]: I0317 15:36:56.426300 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:56 crc kubenswrapper[4767]: I0317 15:36:56.426351 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:56 crc kubenswrapper[4767]: I0317 15:36:56.426365 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:56 crc kubenswrapper[4767]: I0317 15:36:56.426580 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:56 crc kubenswrapper[4767]: I0317 15:36:56.426625 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:56 crc kubenswrapper[4767]: I0317 15:36:56.426639 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:57 crc kubenswrapper[4767]: I0317 15:36:57.770255 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:57 crc kubenswrapper[4767]: I0317 15:36:57.770525 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:57 crc kubenswrapper[4767]: I0317 15:36:57.772619 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:57 crc kubenswrapper[4767]: I0317 15:36:57.772676 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:57 crc kubenswrapper[4767]: I0317 15:36:57.772699 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.104043 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.323091 4767 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.323254 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.385935 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.386205 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.387488 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.387635 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.387672 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.430840 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.432082 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.432119 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:36:58 crc kubenswrapper[4767]: I0317 15:36:58.432131 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:36:59 crc kubenswrapper[4767]: E0317 15:36:59.458812 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.306623 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 17 15:37:03 crc kubenswrapper[4767]: W0317 15:37:03.331191 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.331303 4767 trace.go:236] Trace[360659180]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Mar-2026 15:36:53.329) (total time: 10001ms): Mar 17 15:37:03 crc kubenswrapper[4767]: Trace[360659180]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (15:37:03.331) Mar 17 15:37:03 crc kubenswrapper[4767]: Trace[360659180]: [10.001539971s] [10.001539971s] END Mar 17 15:37:03 crc kubenswrapper[4767]: E0317 15:37:03.331332 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.427779 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.428258 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.430090 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.430153 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.430195 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.474657 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.475055 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.476387 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.476425 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.476434 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.491802 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 17 15:37:03 crc kubenswrapper[4767]: E0317 15:37:03.738296 4767 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:03Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:03 crc kubenswrapper[4767]: E0317 15:37:03.739754 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:03Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.740479 4767 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.740523 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 17 15:37:03 crc kubenswrapper[4767]: E0317 15:37:03.741318 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:03Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189daaf7bf416d9e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,LastTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:03 crc kubenswrapper[4767]: W0317 15:37:03.742981 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:03Z is after 2026-02-23T05:33:13Z Mar 17 15:37:03 crc kubenswrapper[4767]: E0317 15:37:03.743052 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:03Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:03 crc kubenswrapper[4767]: W0317 15:37:03.744890 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:03Z is after 2026-02-23T05:33:13Z Mar 17 15:37:03 crc kubenswrapper[4767]: E0317 15:37:03.744955 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:03Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:03 crc kubenswrapper[4767]: E0317 15:37:03.748103 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:03Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.749076 4767 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 15:37:03 crc kubenswrapper[4767]: I0317 15:37:03.749117 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 17 15:37:03 crc kubenswrapper[4767]: W0317 15:37:03.753396 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:03Z is after 2026-02-23T05:33:13Z Mar 17 15:37:03 crc kubenswrapper[4767]: E0317 15:37:03.753506 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:03Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.310990 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:04Z is after 2026-02-23T05:33:13Z Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.453040 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.453787 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.456497 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c214a1fd1674cbd43fd386473a8e997ca090b74b2e6b7cec66393dafdd058d6f" exitCode=255 Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.456618 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c214a1fd1674cbd43fd386473a8e997ca090b74b2e6b7cec66393dafdd058d6f"} Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.456728 4767 scope.go:117] "RemoveContainer" containerID="59588a28aef91c73a40b25c5a32c3fc6e7bdd42fcc66c9fb79efd9165c68db71" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.456740 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.456968 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.458375 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.458412 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.458426 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.458645 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.458752 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.458852 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:04 crc kubenswrapper[4767]: I0317 15:37:04.458986 4767 scope.go:117] "RemoveContainer" containerID="c214a1fd1674cbd43fd386473a8e997ca090b74b2e6b7cec66393dafdd058d6f" Mar 17 15:37:04 crc kubenswrapper[4767]: E0317 15:37:04.459227 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:37:05 crc kubenswrapper[4767]: I0317 15:37:05.102917 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:37:05 crc kubenswrapper[4767]: I0317 15:37:05.311273 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:05Z is after 2026-02-23T05:33:13Z Mar 17 15:37:05 crc kubenswrapper[4767]: I0317 15:37:05.463098 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 17 15:37:05 crc kubenswrapper[4767]: I0317 15:37:05.466282 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:05 crc kubenswrapper[4767]: I0317 15:37:05.467357 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:05 crc kubenswrapper[4767]: I0317 15:37:05.467416 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:05 crc kubenswrapper[4767]: I0317 15:37:05.467432 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:05 crc kubenswrapper[4767]: I0317 15:37:05.468108 4767 scope.go:117] "RemoveContainer" containerID="c214a1fd1674cbd43fd386473a8e997ca090b74b2e6b7cec66393dafdd058d6f" Mar 17 15:37:05 crc kubenswrapper[4767]: E0317 15:37:05.468342 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:37:06 crc kubenswrapper[4767]: I0317 15:37:06.310665 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:06Z is after 2026-02-23T05:33:13Z Mar 17 15:37:07 crc kubenswrapper[4767]: I0317 15:37:07.309946 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:07Z is after 2026-02-23T05:33:13Z Mar 17 15:37:07 crc kubenswrapper[4767]: I0317 15:37:07.777898 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:37:07 crc kubenswrapper[4767]: I0317 15:37:07.778461 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:07 crc kubenswrapper[4767]: I0317 15:37:07.780130 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:07 crc kubenswrapper[4767]: I0317 15:37:07.780483 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:07 crc kubenswrapper[4767]: I0317 15:37:07.780673 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:07 crc kubenswrapper[4767]: I0317 15:37:07.781719 4767 scope.go:117] "RemoveContainer" containerID="c214a1fd1674cbd43fd386473a8e997ca090b74b2e6b7cec66393dafdd058d6f" Mar 17 15:37:07 crc kubenswrapper[4767]: E0317 15:37:07.782128 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:37:07 crc kubenswrapper[4767]: I0317 15:37:07.786072 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:37:08 crc kubenswrapper[4767]: W0317 15:37:08.228642 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:08Z is after 2026-02-23T05:33:13Z Mar 17 15:37:08 crc kubenswrapper[4767]: E0317 15:37:08.228750 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:08Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:08 crc kubenswrapper[4767]: I0317 15:37:08.312703 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:08Z is after 2026-02-23T05:33:13Z Mar 17 15:37:08 crc kubenswrapper[4767]: I0317 15:37:08.323920 4767 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:37:08 crc kubenswrapper[4767]: I0317 15:37:08.324276 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 15:37:08 crc kubenswrapper[4767]: I0317 15:37:08.474860 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:08 crc kubenswrapper[4767]: I0317 15:37:08.476160 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:08 crc kubenswrapper[4767]: I0317 15:37:08.476211 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:08 crc kubenswrapper[4767]: I0317 15:37:08.476223 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:08 crc kubenswrapper[4767]: I0317 15:37:08.476760 4767 scope.go:117] "RemoveContainer" containerID="c214a1fd1674cbd43fd386473a8e997ca090b74b2e6b7cec66393dafdd058d6f" Mar 17 15:37:08 crc kubenswrapper[4767]: E0317 15:37:08.476946 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:37:09 crc kubenswrapper[4767]: I0317 15:37:09.310985 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:09Z is after 2026-02-23T05:33:13Z Mar 17 15:37:09 crc kubenswrapper[4767]: E0317 15:37:09.459794 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 15:37:09 crc kubenswrapper[4767]: W0317 15:37:09.802521 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:09Z is after 2026-02-23T05:33:13Z Mar 17 15:37:09 crc kubenswrapper[4767]: E0317 15:37:09.802610 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:09Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:09 crc kubenswrapper[4767]: I0317 15:37:09.980997 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:37:09 crc kubenswrapper[4767]: I0317 15:37:09.981147 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:09 crc kubenswrapper[4767]: I0317 15:37:09.982242 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:09 crc kubenswrapper[4767]: I0317 15:37:09.982283 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:09 crc kubenswrapper[4767]: I0317 15:37:09.982301 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:10 crc kubenswrapper[4767]: E0317 15:37:10.142779 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:10Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 17 15:37:10 crc kubenswrapper[4767]: I0317 15:37:10.148967 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:10 crc kubenswrapper[4767]: I0317 15:37:10.149884 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:10 crc kubenswrapper[4767]: I0317 15:37:10.149917 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:10 crc kubenswrapper[4767]: I0317 15:37:10.149932 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:10 crc kubenswrapper[4767]: I0317 15:37:10.149951 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:37:10 crc kubenswrapper[4767]: E0317 15:37:10.152386 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:10Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 15:37:10 crc kubenswrapper[4767]: I0317 15:37:10.310382 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:10Z is after 2026-02-23T05:33:13Z Mar 17 15:37:10 crc kubenswrapper[4767]: W0317 15:37:10.401211 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:10Z is after 2026-02-23T05:33:13Z Mar 17 15:37:10 crc kubenswrapper[4767]: E0317 15:37:10.401355 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:10Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:11 crc kubenswrapper[4767]: I0317 15:37:11.311742 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:11Z is after 2026-02-23T05:33:13Z Mar 17 15:37:11 crc kubenswrapper[4767]: I0317 15:37:11.832636 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 15:37:11 crc kubenswrapper[4767]: E0317 15:37:11.838609 4767 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:11Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:11 crc kubenswrapper[4767]: W0317 15:37:11.912810 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:11Z is after 2026-02-23T05:33:13Z Mar 17 15:37:11 crc kubenswrapper[4767]: E0317 15:37:11.913076 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:11Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:12 crc kubenswrapper[4767]: I0317 15:37:12.310847 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:12Z is after 2026-02-23T05:33:13Z Mar 17 15:37:13 crc kubenswrapper[4767]: I0317 15:37:13.313503 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:13Z is after 2026-02-23T05:33:13Z Mar 17 15:37:13 crc kubenswrapper[4767]: E0317 15:37:13.745414 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:13Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189daaf7bf416d9e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,LastTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:14 crc kubenswrapper[4767]: I0317 15:37:14.310982 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:14Z is after 2026-02-23T05:33:13Z Mar 17 15:37:15 crc kubenswrapper[4767]: I0317 15:37:15.309703 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:15Z is after 2026-02-23T05:33:13Z Mar 17 15:37:16 crc kubenswrapper[4767]: I0317 15:37:16.311523 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:16Z is after 2026-02-23T05:33:13Z Mar 17 15:37:17 crc kubenswrapper[4767]: E0317 15:37:17.148376 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:17Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 17 15:37:17 crc kubenswrapper[4767]: I0317 15:37:17.152695 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:17 crc kubenswrapper[4767]: I0317 15:37:17.153955 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:17 crc kubenswrapper[4767]: I0317 15:37:17.153998 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:17 crc kubenswrapper[4767]: I0317 15:37:17.154011 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:17 crc kubenswrapper[4767]: I0317 15:37:17.154036 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:37:17 crc kubenswrapper[4767]: E0317 15:37:17.157512 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:17Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 15:37:17 crc kubenswrapper[4767]: I0317 15:37:17.311783 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:17Z is after 2026-02-23T05:33:13Z Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.310611 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:18Z is after 2026-02-23T05:33:13Z Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.323126 4767 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.323315 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.323412 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.323610 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.325044 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.325074 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.325083 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.325647 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"bc151899c4826bb8672c4ffbc2b1a199e2193f4f5e9fa86f6c68565ff7cf0c68"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.325862 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://bc151899c4826bb8672c4ffbc2b1a199e2193f4f5e9fa86f6c68565ff7cf0c68" gracePeriod=30 Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.504051 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.504537 4767 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="bc151899c4826bb8672c4ffbc2b1a199e2193f4f5e9fa86f6c68565ff7cf0c68" exitCode=255 Mar 17 15:37:18 crc kubenswrapper[4767]: I0317 15:37:18.504592 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"bc151899c4826bb8672c4ffbc2b1a199e2193f4f5e9fa86f6c68565ff7cf0c68"} Mar 17 15:37:18 crc kubenswrapper[4767]: W0317 15:37:18.904320 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:18Z is after 2026-02-23T05:33:13Z Mar 17 15:37:18 crc kubenswrapper[4767]: E0317 15:37:18.904389 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:18Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:19 crc kubenswrapper[4767]: I0317 15:37:19.308425 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:19Z is after 2026-02-23T05:33:13Z Mar 17 15:37:19 crc kubenswrapper[4767]: E0317 15:37:19.460003 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 15:37:19 crc kubenswrapper[4767]: I0317 15:37:19.512114 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 17 15:37:19 crc kubenswrapper[4767]: I0317 15:37:19.512627 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"805aac1aa5eda8c21499dd0ab570c3d30f25303a39357979f5314a0f602a0eb6"} Mar 17 15:37:19 crc kubenswrapper[4767]: I0317 15:37:19.512804 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:19 crc kubenswrapper[4767]: I0317 15:37:19.513918 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:19 crc kubenswrapper[4767]: I0317 15:37:19.513976 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:19 crc kubenswrapper[4767]: I0317 15:37:19.513999 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:20 crc kubenswrapper[4767]: W0317 15:37:20.096831 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:20Z is after 2026-02-23T05:33:13Z Mar 17 15:37:20 crc kubenswrapper[4767]: E0317 15:37:20.096939 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:20Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:20 crc kubenswrapper[4767]: I0317 15:37:20.309737 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:20Z is after 2026-02-23T05:33:13Z Mar 17 15:37:20 crc kubenswrapper[4767]: I0317 15:37:20.516442 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:20 crc kubenswrapper[4767]: I0317 15:37:20.517568 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:20 crc kubenswrapper[4767]: I0317 15:37:20.517620 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:20 crc kubenswrapper[4767]: I0317 15:37:20.517632 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:21 crc kubenswrapper[4767]: I0317 15:37:21.310952 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:21Z is after 2026-02-23T05:33:13Z Mar 17 15:37:21 crc kubenswrapper[4767]: I0317 15:37:21.354850 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:21 crc kubenswrapper[4767]: I0317 15:37:21.358358 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:21 crc kubenswrapper[4767]: I0317 15:37:21.358429 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:21 crc kubenswrapper[4767]: I0317 15:37:21.358452 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:21 crc kubenswrapper[4767]: I0317 15:37:21.359608 4767 scope.go:117] "RemoveContainer" containerID="c214a1fd1674cbd43fd386473a8e997ca090b74b2e6b7cec66393dafdd058d6f" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.312000 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:22Z is after 2026-02-23T05:33:13Z Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.522147 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.522808 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.524526 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="48238564e409a503532da6971e1200f3fbf759afa7ee87ecb11b3b2a880d8261" exitCode=255 Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.524579 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"48238564e409a503532da6971e1200f3fbf759afa7ee87ecb11b3b2a880d8261"} Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.524630 4767 scope.go:117] "RemoveContainer" containerID="c214a1fd1674cbd43fd386473a8e997ca090b74b2e6b7cec66393dafdd058d6f" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.524752 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.525623 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.525649 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.525658 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.526251 4767 scope.go:117] "RemoveContainer" containerID="48238564e409a503532da6971e1200f3fbf759afa7ee87ecb11b3b2a880d8261" Mar 17 15:37:22 crc kubenswrapper[4767]: E0317 15:37:22.526525 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.783989 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.784309 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.786080 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.786151 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:22 crc kubenswrapper[4767]: I0317 15:37:22.786213 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:23 crc kubenswrapper[4767]: I0317 15:37:23.314371 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:23Z is after 2026-02-23T05:33:13Z Mar 17 15:37:23 crc kubenswrapper[4767]: I0317 15:37:23.529241 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 17 15:37:23 crc kubenswrapper[4767]: E0317 15:37:23.751892 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:23Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189daaf7bf416d9e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,LastTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:24 crc kubenswrapper[4767]: E0317 15:37:24.154818 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:24Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 17 15:37:24 crc kubenswrapper[4767]: I0317 15:37:24.157911 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:24 crc kubenswrapper[4767]: I0317 15:37:24.159847 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:24 crc kubenswrapper[4767]: I0317 15:37:24.159903 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:24 crc kubenswrapper[4767]: I0317 15:37:24.159928 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:24 crc kubenswrapper[4767]: I0317 15:37:24.159969 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:37:24 crc kubenswrapper[4767]: E0317 15:37:24.165629 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:24Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 15:37:24 crc kubenswrapper[4767]: I0317 15:37:24.311837 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:24Z is after 2026-02-23T05:33:13Z Mar 17 15:37:24 crc kubenswrapper[4767]: W0317 15:37:24.389465 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:24Z is after 2026-02-23T05:33:13Z Mar 17 15:37:24 crc kubenswrapper[4767]: E0317 15:37:24.389533 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:24Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.103116 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.103317 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.104622 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.104669 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.104683 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.105294 4767 scope.go:117] "RemoveContainer" containerID="48238564e409a503532da6971e1200f3fbf759afa7ee87ecb11b3b2a880d8261" Mar 17 15:37:25 crc kubenswrapper[4767]: E0317 15:37:25.105499 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.309973 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:25Z is after 2026-02-23T05:33:13Z Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.323260 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.323482 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.325154 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.325229 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:25 crc kubenswrapper[4767]: I0317 15:37:25.325262 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:26 crc kubenswrapper[4767]: I0317 15:37:26.310807 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:26Z is after 2026-02-23T05:33:13Z Mar 17 15:37:27 crc kubenswrapper[4767]: I0317 15:37:27.312041 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:27Z is after 2026-02-23T05:33:13Z Mar 17 15:37:27 crc kubenswrapper[4767]: W0317 15:37:27.643910 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:27Z is after 2026-02-23T05:33:13Z Mar 17 15:37:27 crc kubenswrapper[4767]: E0317 15:37:27.644020 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:27Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:27 crc kubenswrapper[4767]: I0317 15:37:27.886267 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 15:37:27 crc kubenswrapper[4767]: E0317 15:37:27.890491 4767 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:27Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:27 crc kubenswrapper[4767]: E0317 15:37:27.891800 4767 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 17 15:37:28 crc kubenswrapper[4767]: I0317 15:37:28.310624 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:28Z is after 2026-02-23T05:33:13Z Mar 17 15:37:28 crc kubenswrapper[4767]: I0317 15:37:28.323767 4767 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:37:28 crc kubenswrapper[4767]: I0317 15:37:28.323904 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 15:37:29 crc kubenswrapper[4767]: I0317 15:37:29.310148 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:29Z is after 2026-02-23T05:33:13Z Mar 17 15:37:29 crc kubenswrapper[4767]: E0317 15:37:29.460157 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 15:37:30 crc kubenswrapper[4767]: I0317 15:37:30.309652 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:30Z is after 2026-02-23T05:33:13Z Mar 17 15:37:30 crc kubenswrapper[4767]: I0317 15:37:30.457056 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:37:30 crc kubenswrapper[4767]: I0317 15:37:30.457451 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:30 crc kubenswrapper[4767]: I0317 15:37:30.458467 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:30 crc kubenswrapper[4767]: I0317 15:37:30.458520 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:30 crc kubenswrapper[4767]: I0317 15:37:30.458534 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:30 crc kubenswrapper[4767]: I0317 15:37:30.459229 4767 scope.go:117] "RemoveContainer" containerID="48238564e409a503532da6971e1200f3fbf759afa7ee87ecb11b3b2a880d8261" Mar 17 15:37:30 crc kubenswrapper[4767]: E0317 15:37:30.459457 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:37:31 crc kubenswrapper[4767]: E0317 15:37:31.161067 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:31Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 17 15:37:31 crc kubenswrapper[4767]: I0317 15:37:31.166329 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:31 crc kubenswrapper[4767]: I0317 15:37:31.167920 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:31 crc kubenswrapper[4767]: I0317 15:37:31.168148 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:31 crc kubenswrapper[4767]: I0317 15:37:31.168402 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:31 crc kubenswrapper[4767]: I0317 15:37:31.168609 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:37:31 crc kubenswrapper[4767]: E0317 15:37:31.172545 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:31Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 15:37:31 crc kubenswrapper[4767]: I0317 15:37:31.311125 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:31Z is after 2026-02-23T05:33:13Z Mar 17 15:37:32 crc kubenswrapper[4767]: I0317 15:37:32.311516 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:32Z is after 2026-02-23T05:33:13Z Mar 17 15:37:33 crc kubenswrapper[4767]: I0317 15:37:33.311506 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:33Z is after 2026-02-23T05:33:13Z Mar 17 15:37:33 crc kubenswrapper[4767]: E0317 15:37:33.756445 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:33Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189daaf7bf416d9e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,LastTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:34 crc kubenswrapper[4767]: I0317 15:37:34.310091 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:34Z is after 2026-02-23T05:33:13Z Mar 17 15:37:35 crc kubenswrapper[4767]: I0317 15:37:35.313221 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:35Z is after 2026-02-23T05:33:13Z Mar 17 15:37:35 crc kubenswrapper[4767]: W0317 15:37:35.362249 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:35Z is after 2026-02-23T05:33:13Z Mar 17 15:37:35 crc kubenswrapper[4767]: E0317 15:37:35.362685 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:35Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:36 crc kubenswrapper[4767]: I0317 15:37:36.311856 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:36Z is after 2026-02-23T05:33:13Z Mar 17 15:37:37 crc kubenswrapper[4767]: I0317 15:37:37.312320 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:37Z is after 2026-02-23T05:33:13Z Mar 17 15:37:38 crc kubenswrapper[4767]: E0317 15:37:38.165597 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:38Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 17 15:37:38 crc kubenswrapper[4767]: I0317 15:37:38.173251 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:38 crc kubenswrapper[4767]: I0317 15:37:38.174640 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:38 crc kubenswrapper[4767]: I0317 15:37:38.174845 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:38 crc kubenswrapper[4767]: I0317 15:37:38.174987 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:38 crc kubenswrapper[4767]: I0317 15:37:38.175228 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:37:38 crc kubenswrapper[4767]: E0317 15:37:38.178300 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:38Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 15:37:38 crc kubenswrapper[4767]: I0317 15:37:38.311676 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:38Z is after 2026-02-23T05:33:13Z Mar 17 15:37:38 crc kubenswrapper[4767]: I0317 15:37:38.324072 4767 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:37:38 crc kubenswrapper[4767]: I0317 15:37:38.324135 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 15:37:39 crc kubenswrapper[4767]: I0317 15:37:39.311792 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:39Z is after 2026-02-23T05:33:13Z Mar 17 15:37:39 crc kubenswrapper[4767]: E0317 15:37:39.460385 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 15:37:40 crc kubenswrapper[4767]: I0317 15:37:40.311804 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:40Z is after 2026-02-23T05:33:13Z Mar 17 15:37:40 crc kubenswrapper[4767]: I0317 15:37:40.480081 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 15:37:40 crc kubenswrapper[4767]: I0317 15:37:40.480372 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:40 crc kubenswrapper[4767]: I0317 15:37:40.481907 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:40 crc kubenswrapper[4767]: I0317 15:37:40.481971 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:40 crc kubenswrapper[4767]: I0317 15:37:40.481990 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:41 crc kubenswrapper[4767]: I0317 15:37:41.309954 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:41Z is after 2026-02-23T05:33:13Z Mar 17 15:37:42 crc kubenswrapper[4767]: I0317 15:37:42.309682 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:42Z is after 2026-02-23T05:33:13Z Mar 17 15:37:43 crc kubenswrapper[4767]: I0317 15:37:43.311708 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:43Z is after 2026-02-23T05:33:13Z Mar 17 15:37:43 crc kubenswrapper[4767]: I0317 15:37:43.353428 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:43 crc kubenswrapper[4767]: I0317 15:37:43.354722 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:43 crc kubenswrapper[4767]: I0317 15:37:43.354753 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:43 crc kubenswrapper[4767]: I0317 15:37:43.354765 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:43 crc kubenswrapper[4767]: I0317 15:37:43.355337 4767 scope.go:117] "RemoveContainer" containerID="48238564e409a503532da6971e1200f3fbf759afa7ee87ecb11b3b2a880d8261" Mar 17 15:37:43 crc kubenswrapper[4767]: E0317 15:37:43.761107 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:43Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189daaf7bf416d9e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,LastTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:44 crc kubenswrapper[4767]: I0317 15:37:44.310555 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:13Z Mar 17 15:37:44 crc kubenswrapper[4767]: I0317 15:37:44.590004 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 17 15:37:44 crc kubenswrapper[4767]: I0317 15:37:44.590826 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 17 15:37:44 crc kubenswrapper[4767]: I0317 15:37:44.593415 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8" exitCode=255 Mar 17 15:37:44 crc kubenswrapper[4767]: I0317 15:37:44.593465 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8"} Mar 17 15:37:44 crc kubenswrapper[4767]: I0317 15:37:44.593527 4767 scope.go:117] "RemoveContainer" containerID="48238564e409a503532da6971e1200f3fbf759afa7ee87ecb11b3b2a880d8261" Mar 17 15:37:44 crc kubenswrapper[4767]: I0317 15:37:44.593772 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:44 crc kubenswrapper[4767]: I0317 15:37:44.599033 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:44 crc kubenswrapper[4767]: I0317 15:37:44.599195 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:44 crc kubenswrapper[4767]: I0317 15:37:44.599302 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:44 crc kubenswrapper[4767]: I0317 15:37:44.600051 4767 scope.go:117] "RemoveContainer" containerID="8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8" Mar 17 15:37:44 crc kubenswrapper[4767]: E0317 15:37:44.600366 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:37:44 crc kubenswrapper[4767]: W0317 15:37:44.663810 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:13Z Mar 17 15:37:44 crc kubenswrapper[4767]: E0317 15:37:44.664114 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.102520 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:37:45 crc kubenswrapper[4767]: E0317 15:37:45.171497 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:45Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.178449 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.180082 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.180133 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.180152 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.180215 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:37:45 crc kubenswrapper[4767]: E0317 15:37:45.185423 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:45Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.313211 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.598167 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.602438 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.604200 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.604277 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.604300 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:45 crc kubenswrapper[4767]: I0317 15:37:45.605051 4767 scope.go:117] "RemoveContainer" containerID="8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8" Mar 17 15:37:45 crc kubenswrapper[4767]: E0317 15:37:45.605387 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:37:46 crc kubenswrapper[4767]: I0317 15:37:46.313704 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:47 crc kubenswrapper[4767]: I0317 15:37:47.313482 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.311347 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.323782 4767 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.323886 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.323977 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.324262 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.325908 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.325984 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.326003 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.326823 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"805aac1aa5eda8c21499dd0ab570c3d30f25303a39357979f5314a0f602a0eb6"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.326984 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://805aac1aa5eda8c21499dd0ab570c3d30f25303a39357979f5314a0f602a0eb6" gracePeriod=30 Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.612361 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.613602 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.613980 4767 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="805aac1aa5eda8c21499dd0ab570c3d30f25303a39357979f5314a0f602a0eb6" exitCode=255 Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.614014 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"805aac1aa5eda8c21499dd0ab570c3d30f25303a39357979f5314a0f602a0eb6"} Mar 17 15:37:48 crc kubenswrapper[4767]: I0317 15:37:48.614048 4767 scope.go:117] "RemoveContainer" containerID="bc151899c4826bb8672c4ffbc2b1a199e2193f4f5e9fa86f6c68565ff7cf0c68" Mar 17 15:37:49 crc kubenswrapper[4767]: I0317 15:37:49.313748 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:49 crc kubenswrapper[4767]: E0317 15:37:49.460597 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 15:37:49 crc kubenswrapper[4767]: I0317 15:37:49.619029 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 17 15:37:49 crc kubenswrapper[4767]: I0317 15:37:49.620565 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c60814e9521162ab99ae4c6e4913699ef3fdba53dd73a91044129b556b432c32"} Mar 17 15:37:49 crc kubenswrapper[4767]: I0317 15:37:49.620720 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:49 crc kubenswrapper[4767]: I0317 15:37:49.621855 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:49 crc kubenswrapper[4767]: I0317 15:37:49.621916 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:49 crc kubenswrapper[4767]: I0317 15:37:49.621940 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:50 crc kubenswrapper[4767]: I0317 15:37:50.314505 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:50 crc kubenswrapper[4767]: I0317 15:37:50.456896 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:37:50 crc kubenswrapper[4767]: I0317 15:37:50.457070 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:50 crc kubenswrapper[4767]: I0317 15:37:50.458676 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:50 crc kubenswrapper[4767]: I0317 15:37:50.458786 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:50 crc kubenswrapper[4767]: I0317 15:37:50.458816 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:50 crc kubenswrapper[4767]: I0317 15:37:50.459755 4767 scope.go:117] "RemoveContainer" containerID="8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8" Mar 17 15:37:50 crc kubenswrapper[4767]: E0317 15:37:50.460043 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:37:50 crc kubenswrapper[4767]: I0317 15:37:50.622888 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:50 crc kubenswrapper[4767]: I0317 15:37:50.623922 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:50 crc kubenswrapper[4767]: I0317 15:37:50.623961 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:50 crc kubenswrapper[4767]: I0317 15:37:50.623976 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:51 crc kubenswrapper[4767]: I0317 15:37:51.311282 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:52 crc kubenswrapper[4767]: E0317 15:37:52.178247 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 15:37:52 crc kubenswrapper[4767]: I0317 15:37:52.186747 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:52 crc kubenswrapper[4767]: I0317 15:37:52.189025 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:52 crc kubenswrapper[4767]: I0317 15:37:52.189095 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:52 crc kubenswrapper[4767]: I0317 15:37:52.189121 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:52 crc kubenswrapper[4767]: I0317 15:37:52.189165 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:37:52 crc kubenswrapper[4767]: E0317 15:37:52.196993 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 15:37:52 crc kubenswrapper[4767]: I0317 15:37:52.313812 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:52 crc kubenswrapper[4767]: I0317 15:37:52.784356 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:37:52 crc kubenswrapper[4767]: I0317 15:37:52.784589 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:52 crc kubenswrapper[4767]: I0317 15:37:52.786806 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:52 crc kubenswrapper[4767]: I0317 15:37:52.786970 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:52 crc kubenswrapper[4767]: I0317 15:37:52.787096 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:53 crc kubenswrapper[4767]: I0317 15:37:53.314413 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.765742 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7bf416d9e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,LastTimestamp:2026-03-17 15:36:49.296223646 +0000 UTC m=+0.709539723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.770898 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2aee0ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353728237 +0000 UTC m=+0.767044284,LastTimestamp:2026-03-17 15:36:49.353728237 +0000 UTC m=+0.767044284,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.775138 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2afec6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353796719 +0000 UTC m=+0.767112766,LastTimestamp:2026-03-17 15:36:49.353796719 +0000 UTC m=+0.767112766,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.779877 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2b02379 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353810809 +0000 UTC m=+0.767126856,LastTimestamp:2026-03-17 15:36:49.353810809 +0000 UTC m=+0.767126856,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.783771 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c8845d02 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.45160525 +0000 UTC m=+0.864921307,LastTimestamp:2026-03-17 15:36:49.45160525 +0000 UTC m=+0.864921307,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.788536 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2aee0ed\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2aee0ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353728237 +0000 UTC m=+0.767044284,LastTimestamp:2026-03-17 15:36:49.455264926 +0000 UTC m=+0.868580973,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.792334 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2afec6f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2afec6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353796719 +0000 UTC m=+0.767112766,LastTimestamp:2026-03-17 15:36:49.455291257 +0000 UTC m=+0.868607304,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.796725 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2b02379\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2b02379 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353810809 +0000 UTC m=+0.767126856,LastTimestamp:2026-03-17 15:36:49.455303017 +0000 UTC m=+0.868619064,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.801647 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2aee0ed\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2aee0ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353728237 +0000 UTC m=+0.767044284,LastTimestamp:2026-03-17 15:36:49.456332537 +0000 UTC m=+0.869648604,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.806255 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2afec6f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2afec6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353796719 +0000 UTC m=+0.767112766,LastTimestamp:2026-03-17 15:36:49.456357318 +0000 UTC m=+0.869673375,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.811691 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2b02379\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2b02379 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353810809 +0000 UTC m=+0.767126856,LastTimestamp:2026-03-17 15:36:49.456369418 +0000 UTC m=+0.869685475,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.817809 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2aee0ed\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2aee0ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353728237 +0000 UTC m=+0.767044284,LastTimestamp:2026-03-17 15:36:49.457457169 +0000 UTC m=+0.870773236,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.822871 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2afec6f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2afec6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353796719 +0000 UTC m=+0.767112766,LastTimestamp:2026-03-17 15:36:49.45748414 +0000 UTC m=+0.870800197,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.827314 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2b02379\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2b02379 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353810809 +0000 UTC m=+0.767126856,LastTimestamp:2026-03-17 15:36:49.457497411 +0000 UTC m=+0.870813468,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.831486 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2aee0ed\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2aee0ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353728237 +0000 UTC m=+0.767044284,LastTimestamp:2026-03-17 15:36:49.457565443 +0000 UTC m=+0.870881490,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.834910 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2afec6f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2afec6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353796719 +0000 UTC m=+0.767112766,LastTimestamp:2026-03-17 15:36:49.457586323 +0000 UTC m=+0.870902370,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.838806 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2b02379\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2b02379 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353810809 +0000 UTC m=+0.767126856,LastTimestamp:2026-03-17 15:36:49.457596434 +0000 UTC m=+0.870912481,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.842400 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2aee0ed\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2aee0ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353728237 +0000 UTC m=+0.767044284,LastTimestamp:2026-03-17 15:36:49.45884755 +0000 UTC m=+0.872163597,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.849646 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2afec6f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2afec6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353796719 +0000 UTC m=+0.767112766,LastTimestamp:2026-03-17 15:36:49.45886172 +0000 UTC m=+0.872177767,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.853905 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2b02379\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2b02379 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353810809 +0000 UTC m=+0.767126856,LastTimestamp:2026-03-17 15:36:49.4588716 +0000 UTC m=+0.872187647,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.857857 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2aee0ed\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2aee0ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353728237 +0000 UTC m=+0.767044284,LastTimestamp:2026-03-17 15:36:49.458887531 +0000 UTC m=+0.872203588,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.861154 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2aee0ed\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2aee0ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353728237 +0000 UTC m=+0.767044284,LastTimestamp:2026-03-17 15:36:49.458910872 +0000 UTC m=+0.872226939,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.864610 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2afec6f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2afec6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353796719 +0000 UTC m=+0.767112766,LastTimestamp:2026-03-17 15:36:49.458930742 +0000 UTC m=+0.872246799,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.868378 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2b02379\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2b02379 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353810809 +0000 UTC m=+0.767126856,LastTimestamp:2026-03-17 15:36:49.458949243 +0000 UTC m=+0.872265300,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.872460 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189daaf7c2afec6f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189daaf7c2afec6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.353796719 +0000 UTC m=+0.767112766,LastTimestamp:2026-03-17 15:36:49.459028245 +0000 UTC m=+0.872344322,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.877765 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf7e13709e8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.865968104 +0000 UTC m=+1.279284171,LastTimestamp:2026-03-17 15:36:49.865968104 +0000 UTC m=+1.279284171,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.881943 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189daaf7e1606c5b openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.868680283 +0000 UTC m=+1.281996380,LastTimestamp:2026-03-17 15:36:49.868680283 +0000 UTC m=+1.281996380,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.886053 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf7e18a5d6c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.871428972 +0000 UTC m=+1.284745039,LastTimestamp:2026-03-17 15:36:49.871428972 +0000 UTC m=+1.284745039,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.893914 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf7e2327eb5 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.882447541 +0000 UTC m=+1.295763598,LastTimestamp:2026-03-17 15:36:49.882447541 +0000 UTC m=+1.295763598,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.898736 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf7e2b3a9ea openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:49.890912746 +0000 UTC m=+1.304228803,LastTimestamp:2026-03-17 15:36:49.890912746 +0000 UTC m=+1.304228803,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.903407 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf8042cf583 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.452510083 +0000 UTC m=+1.865826140,LastTimestamp:2026-03-17 15:36:50.452510083 +0000 UTC m=+1.865826140,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.907874 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf80436e7e0 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.453161952 +0000 UTC m=+1.866478009,LastTimestamp:2026-03-17 15:36:50.453161952 +0000 UTC m=+1.866478009,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.912330 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189daaf8043c359c openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.453509532 +0000 UTC m=+1.866825599,LastTimestamp:2026-03-17 15:36:50.453509532 +0000 UTC m=+1.866825599,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.916262 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8043d017b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.453561723 +0000 UTC m=+1.866877790,LastTimestamp:2026-03-17 15:36:50.453561723 +0000 UTC m=+1.866877790,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.920185 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf804da76ee openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.463880942 +0000 UTC m=+1.877197009,LastTimestamp:2026-03-17 15:36:50.463880942 +0000 UTC m=+1.877197009,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.929090 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf804f16ec9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.465386185 +0000 UTC m=+1.878702242,LastTimestamp:2026-03-17 15:36:50.465386185 +0000 UTC m=+1.878702242,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.934404 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189daaf8050a463b openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.467014203 +0000 UTC m=+1.880330250,LastTimestamp:2026-03-17 15:36:50.467014203 +0000 UTC m=+1.880330250,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.939146 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf8050ac1c0 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.467045824 +0000 UTC m=+1.880361891,LastTimestamp:2026-03-17 15:36:50.467045824 +0000 UTC m=+1.880361891,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.942749 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf8055d19a8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.47244228 +0000 UTC m=+1.885758327,LastTimestamp:2026-03-17 15:36:50.47244228 +0000 UTC m=+1.885758327,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.949802 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8058b3885 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.475464837 +0000 UTC m=+1.888780894,LastTimestamp:2026-03-17 15:36:50.475464837 +0000 UTC m=+1.888780894,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.953451 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf805a8fe12 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.477415954 +0000 UTC m=+1.890732011,LastTimestamp:2026-03-17 15:36:50.477415954 +0000 UTC m=+1.890732011,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.957587 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf8177aa283 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.776367747 +0000 UTC m=+2.189683814,LastTimestamp:2026-03-17 15:36:50.776367747 +0000 UTC m=+2.189683814,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.961346 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf81859312b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.790953259 +0000 UTC m=+2.204269346,LastTimestamp:2026-03-17 15:36:50.790953259 +0000 UTC m=+2.204269346,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.965018 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf81879ca91 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.793089681 +0000 UTC m=+2.206405728,LastTimestamp:2026-03-17 15:36:50.793089681 +0000 UTC m=+2.206405728,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.970493 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf824d26e75 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.000225397 +0000 UTC m=+2.413541454,LastTimestamp:2026-03-17 15:36:51.000225397 +0000 UTC m=+2.413541454,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:53 crc kubenswrapper[4767]: E0317 15:37:53.980322 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf825950eb6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.012980406 +0000 UTC m=+2.426296453,LastTimestamp:2026-03-17 15:36:51.012980406 +0000 UTC m=+2.426296453,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.019355 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf825abb2bd openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.014464189 +0000 UTC m=+2.427780236,LastTimestamp:2026-03-17 15:36:51.014464189 +0000 UTC m=+2.427780236,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.024268 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf8319ca720 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.214804768 +0000 UTC m=+2.628120835,LastTimestamp:2026-03-17 15:36:51.214804768 +0000 UTC m=+2.628120835,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.028581 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf833350e3d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.241569853 +0000 UTC m=+2.654885900,LastTimestamp:2026-03-17 15:36:51.241569853 +0000 UTC m=+2.654885900,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.032558 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf83ae7a0f4 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.370713332 +0000 UTC m=+2.784029379,LastTimestamp:2026-03-17 15:36:51.370713332 +0000 UTC m=+2.784029379,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.039116 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf83b19ecaf openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.374009519 +0000 UTC m=+2.787325566,LastTimestamp:2026-03-17 15:36:51.374009519 +0000 UTC m=+2.787325566,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.053103 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189daaf83b8106c8 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.380766408 +0000 UTC m=+2.794082445,LastTimestamp:2026-03-17 15:36:51.380766408 +0000 UTC m=+2.794082445,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.058119 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf83b8126bc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.380774588 +0000 UTC m=+2.794090635,LastTimestamp:2026-03-17 15:36:51.380774588 +0000 UTC m=+2.794090635,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.063588 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf848664f79 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.597119353 +0000 UTC m=+3.010435400,LastTimestamp:2026-03-17 15:36:51.597119353 +0000 UTC m=+3.010435400,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.078978 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189daaf8487e8d6f openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.598708079 +0000 UTC m=+3.012024126,LastTimestamp:2026-03-17 15:36:51.598708079 +0000 UTC m=+3.012024126,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.082550 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf84881b59f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.598914975 +0000 UTC m=+3.012231022,LastTimestamp:2026-03-17 15:36:51.598914975 +0000 UTC m=+3.012231022,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.087445 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf84882b4dd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.598980317 +0000 UTC m=+3.012296364,LastTimestamp:2026-03-17 15:36:51.598980317 +0000 UTC m=+3.012296364,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.091084 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf84980b033 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.615625267 +0000 UTC m=+3.028941314,LastTimestamp:2026-03-17 15:36:51.615625267 +0000 UTC m=+3.028941314,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.096348 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf84995e5a8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.617015208 +0000 UTC m=+3.030331265,LastTimestamp:2026-03-17 15:36:51.617015208 +0000 UTC m=+3.030331265,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.100752 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf849c83d75 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.620314485 +0000 UTC m=+3.033630532,LastTimestamp:2026-03-17 15:36:51.620314485 +0000 UTC m=+3.033630532,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.104850 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf849c91477 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.620369527 +0000 UTC m=+3.033685574,LastTimestamp:2026-03-17 15:36:51.620369527 +0000 UTC m=+3.033685574,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.108233 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf849d4ab4f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.621129039 +0000 UTC m=+3.034445086,LastTimestamp:2026-03-17 15:36:51.621129039 +0000 UTC m=+3.034445086,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.111554 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189daaf849f40323 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.623183139 +0000 UTC m=+3.036499186,LastTimestamp:2026-03-17 15:36:51.623183139 +0000 UTC m=+3.036499186,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.117248 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf854e61ad9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.806821081 +0000 UTC m=+3.220137128,LastTimestamp:2026-03-17 15:36:51.806821081 +0000 UTC m=+3.220137128,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.123239 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf854efd8f2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.80745957 +0000 UTC m=+3.220775617,LastTimestamp:2026-03-17 15:36:51.80745957 +0000 UTC m=+3.220775617,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.127127 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf855eb68d9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.823945945 +0000 UTC m=+3.237261992,LastTimestamp:2026-03-17 15:36:51.823945945 +0000 UTC m=+3.237261992,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.135390 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf855fa095f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.824904543 +0000 UTC m=+3.238220590,LastTimestamp:2026-03-17 15:36:51.824904543 +0000 UTC m=+3.238220590,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.142347 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf85617b7a5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.826849701 +0000 UTC m=+3.240165748,LastTimestamp:2026-03-17 15:36:51.826849701 +0000 UTC m=+3.240165748,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.146019 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf8562283a1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:51.827557281 +0000 UTC m=+3.240873318,LastTimestamp:2026-03-17 15:36:51.827557281 +0000 UTC m=+3.240873318,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.151001 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf8624dc83f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.031719487 +0000 UTC m=+3.445035564,LastTimestamp:2026-03-17 15:36:52.031719487 +0000 UTC m=+3.445035564,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.158059 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf8625243d2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.032013266 +0000 UTC m=+3.445329333,LastTimestamp:2026-03-17 15:36:52.032013266 +0000 UTC m=+3.445329333,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.165230 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf8630d2580 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.044260736 +0000 UTC m=+3.457576773,LastTimestamp:2026-03-17 15:36:52.044260736 +0000 UTC m=+3.457576773,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.168861 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf8631aba34 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.045150772 +0000 UTC m=+3.458466819,LastTimestamp:2026-03-17 15:36:52.045150772 +0000 UTC m=+3.458466819,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.175153 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189daaf8635481f8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.048937464 +0000 UTC m=+3.462253511,LastTimestamp:2026-03-17 15:36:52.048937464 +0000 UTC m=+3.462253511,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.180195 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf86cc131c9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.207055305 +0000 UTC m=+3.620371372,LastTimestamp:2026-03-17 15:36:52.207055305 +0000 UTC m=+3.620371372,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.201070 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf86d9c15d9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.221400537 +0000 UTC m=+3.634716594,LastTimestamp:2026-03-17 15:36:52.221400537 +0000 UTC m=+3.634716594,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.293590 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf86db0c97f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.222757247 +0000 UTC m=+3.636073294,LastTimestamp:2026-03-17 15:36:52.222757247 +0000 UTC m=+3.636073294,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.305577 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8776ff9f9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.386281977 +0000 UTC m=+3.799598024,LastTimestamp:2026-03-17 15:36:52.386281977 +0000 UTC m=+3.799598024,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: I0317 15:37:54.310868 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.312130 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf879f65742 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.428642114 +0000 UTC m=+3.841958161,LastTimestamp:2026-03-17 15:36:52.428642114 +0000 UTC m=+3.841958161,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.313016 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf87aecbd3f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.444790079 +0000 UTC m=+3.858106126,LastTimestamp:2026-03-17 15:36:52.444790079 +0000 UTC m=+3.858106126,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.317288 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8825f6680 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.569745024 +0000 UTC m=+3.983061071,LastTimestamp:2026-03-17 15:36:52.569745024 +0000 UTC m=+3.983061071,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.324112 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf883acf7df openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.591605727 +0000 UTC m=+4.004921774,LastTimestamp:2026-03-17 15:36:52.591605727 +0000 UTC m=+4.004921774,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.328398 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8b3b3dae5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:53.397363429 +0000 UTC m=+4.810679516,LastTimestamp:2026-03-17 15:36:53.397363429 +0000 UTC m=+4.810679516,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.334271 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189daaf86db0c97f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf86db0c97f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.222757247 +0000 UTC m=+3.636073294,LastTimestamp:2026-03-17 15:36:53.406957382 +0000 UTC m=+4.820273449,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.341261 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189daaf879f65742\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf879f65742 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.428642114 +0000 UTC m=+3.841958161,LastTimestamp:2026-03-17 15:36:53.607846191 +0000 UTC m=+5.021162248,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.344329 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8c059efd0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:53.609574352 +0000 UTC m=+5.022890419,LastTimestamp:2026-03-17 15:36:53.609574352 +0000 UTC m=+5.022890419,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.347641 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189daaf87aecbd3f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daaf87aecbd3f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:52.444790079 +0000 UTC m=+3.858106126,LastTimestamp:2026-03-17 15:36:53.618434943 +0000 UTC m=+5.031751030,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.350850 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8c17be651 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:53.628577361 +0000 UTC m=+5.041893438,LastTimestamp:2026-03-17 15:36:53.628577361 +0000 UTC m=+5.041893438,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.354521 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8c18b2f8e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:53.62957915 +0000 UTC m=+5.042895237,LastTimestamp:2026-03-17 15:36:53.62957915 +0000 UTC m=+5.042895237,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.357792 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8cd412356 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:53.82605295 +0000 UTC m=+5.239369017,LastTimestamp:2026-03-17 15:36:53.82605295 +0000 UTC m=+5.239369017,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.361345 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8cddbe88b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:53.836195979 +0000 UTC m=+5.249512036,LastTimestamp:2026-03-17 15:36:53.836195979 +0000 UTC m=+5.249512036,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.364217 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8cded9ee5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:53.837356773 +0000 UTC m=+5.250672820,LastTimestamp:2026-03-17 15:36:53.837356773 +0000 UTC m=+5.250672820,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.367268 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8d8f9af8d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:54.022696845 +0000 UTC m=+5.436012892,LastTimestamp:2026-03-17 15:36:54.022696845 +0000 UTC m=+5.436012892,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.370241 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8da58c816 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:54.045706262 +0000 UTC m=+5.459022319,LastTimestamp:2026-03-17 15:36:54.045706262 +0000 UTC m=+5.459022319,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.373252 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8da673f98 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:54.04665436 +0000 UTC m=+5.459970427,LastTimestamp:2026-03-17 15:36:54.04665436 +0000 UTC m=+5.459970427,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.376506 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8e5b26db7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:54.236130743 +0000 UTC m=+5.649446790,LastTimestamp:2026-03-17 15:36:54.236130743 +0000 UTC m=+5.649446790,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.379672 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8e6b050ad openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:54.252769453 +0000 UTC m=+5.666085500,LastTimestamp:2026-03-17 15:36:54.252769453 +0000 UTC m=+5.666085500,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.383776 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8e6bcabc7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:54.253579207 +0000 UTC m=+5.666895254,LastTimestamp:2026-03-17 15:36:54.253579207 +0000 UTC m=+5.666895254,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.387673 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8f235c0b7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:54.446063799 +0000 UTC m=+5.859379856,LastTimestamp:2026-03-17 15:36:54.446063799 +0000 UTC m=+5.859379856,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.390480 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189daaf8f304d22a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:54.459634218 +0000 UTC m=+5.872950265,LastTimestamp:2026-03-17 15:36:54.459634218 +0000 UTC m=+5.872950265,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.394284 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 15:37:54 crc kubenswrapper[4767]: &Event{ObjectMeta:{kube-controller-manager-crc.189daaf9d94e7a28 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 17 15:37:54 crc kubenswrapper[4767]: body: Mar 17 15:37:54 crc kubenswrapper[4767]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:58.323221032 +0000 UTC m=+9.736537119,LastTimestamp:2026-03-17 15:36:58.323221032 +0000 UTC m=+9.736537119,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 15:37:54 crc kubenswrapper[4767]: > Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.397011 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf9d94fb4f8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:58.323301624 +0000 UTC m=+9.736617711,LastTimestamp:2026-03-17 15:36:58.323301624 +0000 UTC m=+9.736617711,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.400588 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 17 15:37:54 crc kubenswrapper[4767]: &Event{ObjectMeta:{kube-apiserver-crc.189daafb1c33ba38 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 17 15:37:54 crc kubenswrapper[4767]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 15:37:54 crc kubenswrapper[4767]: Mar 17 15:37:54 crc kubenswrapper[4767]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:37:03.740508728 +0000 UTC m=+15.153824775,LastTimestamp:2026-03-17 15:37:03.740508728 +0000 UTC m=+15.153824775,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 15:37:54 crc kubenswrapper[4767]: > Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.406779 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daafb1c3445cb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:37:03.740544459 +0000 UTC m=+15.153860506,LastTimestamp:2026-03-17 15:37:03.740544459 +0000 UTC m=+15.153860506,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.412895 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189daafb1c33ba38\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 17 15:37:54 crc kubenswrapper[4767]: &Event{ObjectMeta:{kube-apiserver-crc.189daafb1c33ba38 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 17 15:37:54 crc kubenswrapper[4767]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 15:37:54 crc kubenswrapper[4767]: Mar 17 15:37:54 crc kubenswrapper[4767]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:37:03.740508728 +0000 UTC m=+15.153824775,LastTimestamp:2026-03-17 15:37:03.749105171 +0000 UTC m=+15.162421218,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 15:37:54 crc kubenswrapper[4767]: > Mar 17 15:37:54 crc kubenswrapper[4767]: W0317 15:37:54.536582 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.536624 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.537157 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189daafb1c3445cb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189daafb1c3445cb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:37:03.740544459 +0000 UTC m=+15.153860506,LastTimestamp:2026-03-17 15:37:03.749136572 +0000 UTC m=+15.162452619,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.551893 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189daaf9d94e7a28\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 15:37:54 crc kubenswrapper[4767]: &Event{ObjectMeta:{kube-controller-manager-crc.189daaf9d94e7a28 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 17 15:37:54 crc kubenswrapper[4767]: body: Mar 17 15:37:54 crc kubenswrapper[4767]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:58.323221032 +0000 UTC m=+9.736537119,LastTimestamp:2026-03-17 15:37:08.324236007 +0000 UTC m=+19.737552094,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 15:37:54 crc kubenswrapper[4767]: > Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.563256 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189daaf9d94fb4f8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf9d94fb4f8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:58.323301624 +0000 UTC m=+9.736617711,LastTimestamp:2026-03-17 15:37:08.324450884 +0000 UTC m=+19.737766971,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.583647 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189daaf9d94e7a28\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 15:37:54 crc kubenswrapper[4767]: &Event{ObjectMeta:{kube-controller-manager-crc.189daaf9d94e7a28 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 17 15:37:54 crc kubenswrapper[4767]: body: Mar 17 15:37:54 crc kubenswrapper[4767]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:58.323221032 +0000 UTC m=+9.736537119,LastTimestamp:2026-03-17 15:37:18.323274455 +0000 UTC m=+29.736590492,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 15:37:54 crc kubenswrapper[4767]: > Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.597230 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189daaf9d94fb4f8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf9d94fb4f8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:58.323301624 +0000 UTC m=+9.736617711,LastTimestamp:2026-03-17 15:37:18.323367918 +0000 UTC m=+29.736683965,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.613138 4767 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daafe818e4eb9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:37:18.325845689 +0000 UTC m=+29.739161736,LastTimestamp:2026-03-17 15:37:18.325845689 +0000 UTC m=+29.739161736,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.625461 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189daaf8050ac1c0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf8050ac1c0 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.467045824 +0000 UTC m=+1.880361891,LastTimestamp:2026-03-17 15:37:18.453786273 +0000 UTC m=+29.867102330,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.630439 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189daaf8177aa283\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf8177aa283 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.776367747 +0000 UTC m=+2.189683814,LastTimestamp:2026-03-17 15:37:18.68153793 +0000 UTC m=+30.094853987,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.649806 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189daaf81859312b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf81859312b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:50.790953259 +0000 UTC m=+2.204269346,LastTimestamp:2026-03-17 15:37:18.691846897 +0000 UTC m=+30.105162954,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.687761 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189daaf9d94e7a28\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 15:37:54 crc kubenswrapper[4767]: &Event{ObjectMeta:{kube-controller-manager-crc.189daaf9d94e7a28 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 17 15:37:54 crc kubenswrapper[4767]: body: Mar 17 15:37:54 crc kubenswrapper[4767]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:58.323221032 +0000 UTC m=+9.736537119,LastTimestamp:2026-03-17 15:37:28.323862715 +0000 UTC m=+39.737178772,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 15:37:54 crc kubenswrapper[4767]: > Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.693484 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189daaf9d94fb4f8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189daaf9d94fb4f8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:58.323301624 +0000 UTC m=+9.736617711,LastTimestamp:2026-03-17 15:37:28.323933797 +0000 UTC m=+39.737249854,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:37:54 crc kubenswrapper[4767]: E0317 15:37:54.698085 4767 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189daaf9d94e7a28\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 15:37:54 crc kubenswrapper[4767]: &Event{ObjectMeta:{kube-controller-manager-crc.189daaf9d94e7a28 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 17 15:37:54 crc kubenswrapper[4767]: body: Mar 17 15:37:54 crc kubenswrapper[4767]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:36:58.323221032 +0000 UTC m=+9.736537119,LastTimestamp:2026-03-17 15:37:38.324111855 +0000 UTC m=+49.737427942,Count:5,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 15:37:54 crc kubenswrapper[4767]: > Mar 17 15:37:55 crc kubenswrapper[4767]: I0317 15:37:55.310669 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:55 crc kubenswrapper[4767]: I0317 15:37:55.323759 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:37:55 crc kubenswrapper[4767]: I0317 15:37:55.323911 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:55 crc kubenswrapper[4767]: I0317 15:37:55.325123 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:55 crc kubenswrapper[4767]: I0317 15:37:55.325198 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:55 crc kubenswrapper[4767]: I0317 15:37:55.325211 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:55 crc kubenswrapper[4767]: I0317 15:37:55.327689 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:37:55 crc kubenswrapper[4767]: I0317 15:37:55.639550 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:55 crc kubenswrapper[4767]: I0317 15:37:55.640275 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:55 crc kubenswrapper[4767]: I0317 15:37:55.640301 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:55 crc kubenswrapper[4767]: I0317 15:37:55.640310 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:56 crc kubenswrapper[4767]: I0317 15:37:56.310217 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:57 crc kubenswrapper[4767]: I0317 15:37:57.311078 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:58 crc kubenswrapper[4767]: I0317 15:37:58.310365 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:59 crc kubenswrapper[4767]: E0317 15:37:59.182440 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 15:37:59 crc kubenswrapper[4767]: I0317 15:37:59.197575 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:37:59 crc kubenswrapper[4767]: I0317 15:37:59.198708 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:37:59 crc kubenswrapper[4767]: I0317 15:37:59.198740 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:37:59 crc kubenswrapper[4767]: I0317 15:37:59.198751 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:37:59 crc kubenswrapper[4767]: I0317 15:37:59.198775 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:37:59 crc kubenswrapper[4767]: E0317 15:37:59.202623 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 15:37:59 crc kubenswrapper[4767]: I0317 15:37:59.310344 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:37:59 crc kubenswrapper[4767]: E0317 15:37:59.460879 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 15:37:59 crc kubenswrapper[4767]: I0317 15:37:59.893311 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 15:37:59 crc kubenswrapper[4767]: I0317 15:37:59.907950 4767 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 17 15:38:00 crc kubenswrapper[4767]: I0317 15:38:00.310593 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:38:01 crc kubenswrapper[4767]: I0317 15:38:01.309848 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:38:01 crc kubenswrapper[4767]: I0317 15:38:01.353684 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:38:01 crc kubenswrapper[4767]: I0317 15:38:01.355222 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:01 crc kubenswrapper[4767]: I0317 15:38:01.355254 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:01 crc kubenswrapper[4767]: I0317 15:38:01.355262 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:02 crc kubenswrapper[4767]: I0317 15:38:02.310147 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:38:02 crc kubenswrapper[4767]: I0317 15:38:02.788822 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:38:02 crc kubenswrapper[4767]: I0317 15:38:02.788952 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:38:02 crc kubenswrapper[4767]: I0317 15:38:02.790065 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:02 crc kubenswrapper[4767]: I0317 15:38:02.790124 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:02 crc kubenswrapper[4767]: I0317 15:38:02.790139 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:03 crc kubenswrapper[4767]: I0317 15:38:03.312886 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:38:04 crc kubenswrapper[4767]: I0317 15:38:04.312694 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:38:04 crc kubenswrapper[4767]: I0317 15:38:04.353875 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:38:04 crc kubenswrapper[4767]: I0317 15:38:04.355586 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:04 crc kubenswrapper[4767]: I0317 15:38:04.356037 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:04 crc kubenswrapper[4767]: I0317 15:38:04.356331 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:04 crc kubenswrapper[4767]: I0317 15:38:04.357832 4767 scope.go:117] "RemoveContainer" containerID="8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8" Mar 17 15:38:04 crc kubenswrapper[4767]: E0317 15:38:04.358610 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:38:05 crc kubenswrapper[4767]: I0317 15:38:05.313288 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:38:06 crc kubenswrapper[4767]: E0317 15:38:06.189663 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 15:38:06 crc kubenswrapper[4767]: I0317 15:38:06.203293 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:38:06 crc kubenswrapper[4767]: I0317 15:38:06.204626 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:06 crc kubenswrapper[4767]: I0317 15:38:06.204663 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:06 crc kubenswrapper[4767]: I0317 15:38:06.204676 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:06 crc kubenswrapper[4767]: I0317 15:38:06.204745 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:38:06 crc kubenswrapper[4767]: E0317 15:38:06.210153 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 15:38:06 crc kubenswrapper[4767]: I0317 15:38:06.311443 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 15:38:06 crc kubenswrapper[4767]: I0317 15:38:06.563388 4767 csr.go:261] certificate signing request csr-nc9p4 is approved, waiting to be issued Mar 17 15:38:06 crc kubenswrapper[4767]: I0317 15:38:06.576252 4767 csr.go:257] certificate signing request csr-nc9p4 is issued Mar 17 15:38:06 crc kubenswrapper[4767]: I0317 15:38:06.681542 4767 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 17 15:38:07 crc kubenswrapper[4767]: I0317 15:38:07.149523 4767 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 17 15:38:07 crc kubenswrapper[4767]: I0317 15:38:07.578218 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-30 02:10:49.080706691 +0000 UTC Mar 17 15:38:07 crc kubenswrapper[4767]: I0317 15:38:07.578284 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6898h32m41.502427693s for next certificate rotation Mar 17 15:38:09 crc kubenswrapper[4767]: E0317 15:38:09.461066 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 15:38:11 crc kubenswrapper[4767]: I0317 15:38:11.785998 4767 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.210500 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.212631 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.212704 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.212730 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.213214 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.225748 4767 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.226291 4767 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.226346 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.232474 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.232544 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.232558 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.232580 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.232595 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:13Z","lastTransitionTime":"2026-03-17T15:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.259367 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.264829 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.264893 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.264919 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.264950 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.264974 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:13Z","lastTransitionTime":"2026-03-17T15:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.281017 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.285985 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.286025 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.286039 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.286059 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.286073 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:13Z","lastTransitionTime":"2026-03-17T15:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.300952 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.305105 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.305152 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.305187 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.305211 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:13 crc kubenswrapper[4767]: I0317 15:38:13.305226 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:13Z","lastTransitionTime":"2026-03-17T15:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.319508 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.319791 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.319833 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.419914 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.520491 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.621608 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.721818 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.822030 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:13 crc kubenswrapper[4767]: E0317 15:38:13.923114 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:14 crc kubenswrapper[4767]: E0317 15:38:14.024250 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:14 crc kubenswrapper[4767]: E0317 15:38:14.125313 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:14 crc kubenswrapper[4767]: E0317 15:38:14.226261 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:14 crc kubenswrapper[4767]: E0317 15:38:14.326365 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:14 crc kubenswrapper[4767]: E0317 15:38:14.426916 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:14 crc kubenswrapper[4767]: E0317 15:38:14.527758 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:14 crc kubenswrapper[4767]: E0317 15:38:14.628128 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:14 crc kubenswrapper[4767]: E0317 15:38:14.728883 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:14 crc kubenswrapper[4767]: E0317 15:38:14.830086 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:14 crc kubenswrapper[4767]: E0317 15:38:14.930507 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:15 crc kubenswrapper[4767]: E0317 15:38:15.031637 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:15 crc kubenswrapper[4767]: E0317 15:38:15.131760 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:15 crc kubenswrapper[4767]: E0317 15:38:15.232955 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:15 crc kubenswrapper[4767]: E0317 15:38:15.333123 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:15 crc kubenswrapper[4767]: E0317 15:38:15.433362 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:15 crc kubenswrapper[4767]: E0317 15:38:15.533974 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:15 crc kubenswrapper[4767]: E0317 15:38:15.634666 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:15 crc kubenswrapper[4767]: E0317 15:38:15.735534 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:15 crc kubenswrapper[4767]: E0317 15:38:15.835652 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:15 crc kubenswrapper[4767]: E0317 15:38:15.936841 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:16 crc kubenswrapper[4767]: E0317 15:38:16.037847 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:16 crc kubenswrapper[4767]: E0317 15:38:16.138907 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:16 crc kubenswrapper[4767]: E0317 15:38:16.239253 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:16 crc kubenswrapper[4767]: E0317 15:38:16.340388 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:16 crc kubenswrapper[4767]: E0317 15:38:16.441374 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:16 crc kubenswrapper[4767]: E0317 15:38:16.541822 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:16 crc kubenswrapper[4767]: E0317 15:38:16.642505 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:16 crc kubenswrapper[4767]: E0317 15:38:16.743163 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:16 crc kubenswrapper[4767]: E0317 15:38:16.844224 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:16 crc kubenswrapper[4767]: E0317 15:38:16.945127 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:17 crc kubenswrapper[4767]: E0317 15:38:17.046142 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:17 crc kubenswrapper[4767]: E0317 15:38:17.147157 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:17 crc kubenswrapper[4767]: E0317 15:38:17.247573 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:17 crc kubenswrapper[4767]: E0317 15:38:17.348220 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:17 crc kubenswrapper[4767]: I0317 15:38:17.353439 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 15:38:17 crc kubenswrapper[4767]: I0317 15:38:17.354593 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:17 crc kubenswrapper[4767]: I0317 15:38:17.354651 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:17 crc kubenswrapper[4767]: I0317 15:38:17.354671 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:17 crc kubenswrapper[4767]: I0317 15:38:17.355668 4767 scope.go:117] "RemoveContainer" containerID="8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8" Mar 17 15:38:17 crc kubenswrapper[4767]: E0317 15:38:17.355976 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 15:38:17 crc kubenswrapper[4767]: E0317 15:38:17.448642 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:17 crc kubenswrapper[4767]: E0317 15:38:17.549748 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:17 crc kubenswrapper[4767]: E0317 15:38:17.649935 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:17 crc kubenswrapper[4767]: E0317 15:38:17.751080 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:17 crc kubenswrapper[4767]: E0317 15:38:17.852095 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:17 crc kubenswrapper[4767]: E0317 15:38:17.953474 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:18 crc kubenswrapper[4767]: E0317 15:38:18.053905 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:18 crc kubenswrapper[4767]: E0317 15:38:18.154858 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:18 crc kubenswrapper[4767]: E0317 15:38:18.255994 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:18 crc kubenswrapper[4767]: E0317 15:38:18.356521 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:18 crc kubenswrapper[4767]: E0317 15:38:18.457726 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:18 crc kubenswrapper[4767]: E0317 15:38:18.559137 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:18 crc kubenswrapper[4767]: E0317 15:38:18.659537 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:18 crc kubenswrapper[4767]: E0317 15:38:18.760412 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:18 crc kubenswrapper[4767]: E0317 15:38:18.861035 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:18 crc kubenswrapper[4767]: E0317 15:38:18.961443 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:19 crc kubenswrapper[4767]: E0317 15:38:19.062558 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:19 crc kubenswrapper[4767]: E0317 15:38:19.163306 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:19 crc kubenswrapper[4767]: E0317 15:38:19.263981 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:19 crc kubenswrapper[4767]: E0317 15:38:19.364375 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:19 crc kubenswrapper[4767]: E0317 15:38:19.461345 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 15:38:19 crc kubenswrapper[4767]: E0317 15:38:19.464560 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:19 crc kubenswrapper[4767]: E0317 15:38:19.565495 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:19 crc kubenswrapper[4767]: E0317 15:38:19.666186 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:19 crc kubenswrapper[4767]: E0317 15:38:19.767271 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:19 crc kubenswrapper[4767]: E0317 15:38:19.867636 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:19 crc kubenswrapper[4767]: E0317 15:38:19.968264 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:20 crc kubenswrapper[4767]: E0317 15:38:20.069393 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:20 crc kubenswrapper[4767]: E0317 15:38:20.171145 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:20 crc kubenswrapper[4767]: E0317 15:38:20.272099 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:20 crc kubenswrapper[4767]: E0317 15:38:20.374244 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:20 crc kubenswrapper[4767]: E0317 15:38:20.475903 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:20 crc kubenswrapper[4767]: E0317 15:38:20.576758 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:20 crc kubenswrapper[4767]: E0317 15:38:20.677060 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:20 crc kubenswrapper[4767]: E0317 15:38:20.777861 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:20 crc kubenswrapper[4767]: E0317 15:38:20.878937 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:20 crc kubenswrapper[4767]: E0317 15:38:20.979902 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:21 crc kubenswrapper[4767]: E0317 15:38:21.080648 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:21 crc kubenswrapper[4767]: E0317 15:38:21.181137 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:21 crc kubenswrapper[4767]: E0317 15:38:21.282073 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:21 crc kubenswrapper[4767]: E0317 15:38:21.383155 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:21 crc kubenswrapper[4767]: E0317 15:38:21.484128 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:21 crc kubenswrapper[4767]: E0317 15:38:21.584787 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:21 crc kubenswrapper[4767]: E0317 15:38:21.685915 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:21 crc kubenswrapper[4767]: E0317 15:38:21.786103 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:21 crc kubenswrapper[4767]: E0317 15:38:21.886872 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:21 crc kubenswrapper[4767]: E0317 15:38:21.987912 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:22 crc kubenswrapper[4767]: E0317 15:38:22.088587 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:22 crc kubenswrapper[4767]: E0317 15:38:22.188707 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:22 crc kubenswrapper[4767]: E0317 15:38:22.289212 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:22 crc kubenswrapper[4767]: E0317 15:38:22.390270 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:22 crc kubenswrapper[4767]: I0317 15:38:22.488234 4767 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 17 15:38:22 crc kubenswrapper[4767]: E0317 15:38:22.490795 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:22 crc kubenswrapper[4767]: E0317 15:38:22.591435 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:22 crc kubenswrapper[4767]: E0317 15:38:22.692253 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:22 crc kubenswrapper[4767]: E0317 15:38:22.793266 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:22 crc kubenswrapper[4767]: E0317 15:38:22.893400 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:22 crc kubenswrapper[4767]: E0317 15:38:22.993792 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:23 crc kubenswrapper[4767]: E0317 15:38:23.096686 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:23 crc kubenswrapper[4767]: E0317 15:38:23.197158 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:23 crc kubenswrapper[4767]: E0317 15:38:23.297429 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:23 crc kubenswrapper[4767]: E0317 15:38:23.398231 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:23 crc kubenswrapper[4767]: E0317 15:38:23.413965 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.418349 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.418401 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.418414 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.418432 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.418449 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:23Z","lastTransitionTime":"2026-03-17T15:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:23 crc kubenswrapper[4767]: E0317 15:38:23.430846 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.434909 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.434941 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.434952 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.434967 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.434979 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:23Z","lastTransitionTime":"2026-03-17T15:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:23 crc kubenswrapper[4767]: E0317 15:38:23.448006 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.452608 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.452688 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.452706 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.452731 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.452749 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:23Z","lastTransitionTime":"2026-03-17T15:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:23 crc kubenswrapper[4767]: E0317 15:38:23.473017 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.478471 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.478518 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.478565 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.478587 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.478602 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:23Z","lastTransitionTime":"2026-03-17T15:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:23 crc kubenswrapper[4767]: E0317 15:38:23.503122 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:23 crc kubenswrapper[4767]: E0317 15:38:23.503353 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 15:38:23 crc kubenswrapper[4767]: E0317 15:38:23.503389 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.570136 4767 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.606468 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.606536 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.606548 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.606568 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.606592 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:23Z","lastTransitionTime":"2026-03-17T15:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.709101 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.709155 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.709171 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.709209 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.709222 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:23Z","lastTransitionTime":"2026-03-17T15:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.811259 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.811287 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.811295 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.811307 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.811316 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:23Z","lastTransitionTime":"2026-03-17T15:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.914020 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.914093 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.914114 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.914140 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:23 crc kubenswrapper[4767]: I0317 15:38:23.914158 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:23Z","lastTransitionTime":"2026-03-17T15:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.016659 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.016706 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.016719 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.016736 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.016754 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:24Z","lastTransitionTime":"2026-03-17T15:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.118992 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.119034 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.119048 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.119064 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.119078 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:24Z","lastTransitionTime":"2026-03-17T15:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.222521 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.222589 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.222653 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.222680 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.222697 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:24Z","lastTransitionTime":"2026-03-17T15:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.326044 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.326102 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.326120 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.326142 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.326161 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:24Z","lastTransitionTime":"2026-03-17T15:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.342282 4767 apiserver.go:52] "Watching apiserver" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.347694 4767 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.347934 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.348439 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.348500 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.348685 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.348913 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.348941 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.348985 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.349011 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.349056 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.349322 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.351598 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.351805 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.351971 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.352111 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.352513 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.352672 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.352866 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.354468 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.355168 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.392732 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.406163 4767 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.411772 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.411829 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.411856 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.411877 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.411892 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.411910 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.411928 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.411944 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.411960 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.411975 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.411989 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412005 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412020 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412035 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412052 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412067 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412084 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412101 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412116 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412131 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412151 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412182 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412224 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412339 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412356 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412373 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412390 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412406 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412421 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412439 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412454 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412470 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412485 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412501 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412519 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412534 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412805 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.412899 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413008 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413043 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413067 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413237 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413348 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413374 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413447 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413508 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413534 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413639 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413765 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.413923 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414068 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414107 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414262 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414378 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414484 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414563 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414590 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414654 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414721 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414795 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414863 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414924 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.414948 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415015 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415076 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415138 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415225 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415287 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415489 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415549 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415612 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415683 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415740 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415799 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415822 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.415919 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.416027 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.416130 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.416586 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.416636 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.416673 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.417437 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.417730 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.417773 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.417776 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.417825 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.417824 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.417991 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.418073 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.418293 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.418313 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.418517 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.418564 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.418664 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.418510 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.418730 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.418919 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.418969 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.419183 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.419251 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.419266 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.419332 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.419410 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.419462 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.419560 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.419583 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.419600 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.419866 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.420744 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.420803 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.417780 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.421863 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.422049 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.422069 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.422147 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.422248 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.422790 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.422839 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.422878 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.422862 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.421370 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.423220 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.423404 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.423712 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.423958 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.424386 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.424474 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.424531 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.424876 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.423080 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.425046 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.425056 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:38:24.921913106 +0000 UTC m=+96.335229193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.425054 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.425106 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.425156 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.425214 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.425244 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.425154 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.423499 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.425670 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.424668 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.425972 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.426384 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.426759 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.426785 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.426947 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.426994 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427056 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427094 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427127 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427133 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427155 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427192 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427452 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427503 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427528 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427537 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427600 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427636 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427666 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427687 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427709 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427731 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427754 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427774 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427796 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427916 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427947 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.427998 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428020 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428041 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428062 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428079 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428184 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428256 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428408 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428408 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428466 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428490 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428556 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428656 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428721 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428460 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428815 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.428877 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.429005 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.429054 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.429087 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.429120 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.429056 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.429455 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.429636 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.429723 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430037 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430128 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430137 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430239 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430337 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430454 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430488 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430544 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430585 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430294 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430821 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430847 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.430876 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.431225 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.431394 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.431406 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.431403 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.431639 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.431715 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.432042 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.432160 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.436444 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.437441 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438575 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438646 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438676 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438698 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438718 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438739 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438767 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438791 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438813 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438828 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438843 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438859 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438873 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438889 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438907 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438928 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438948 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438968 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.438987 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439005 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439022 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439037 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439052 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439069 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439084 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439104 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439136 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439159 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439190 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439238 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439256 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439272 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439288 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439304 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439320 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439336 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439351 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439368 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439386 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439403 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439419 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439436 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439456 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439474 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439496 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439527 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439548 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439568 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439586 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439620 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439641 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439661 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439685 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439704 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439719 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439735 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439752 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439769 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439785 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439800 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439819 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439839 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439860 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439882 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439907 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439932 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.439987 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440016 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440034 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440053 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440070 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440085 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440101 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440115 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440133 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440179 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440225 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440253 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440270 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440289 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440307 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440327 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440373 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.440391 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.441108 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.441854 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.441895 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.441939 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.441976 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442014 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442106 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442121 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442131 4767 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442143 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442157 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442173 4767 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442187 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442214 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442225 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442236 4767 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442248 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442262 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442274 4767 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442286 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442299 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442312 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442325 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442338 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442352 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442364 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442376 4767 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442389 4767 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442401 4767 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442413 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442425 4767 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442437 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442450 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442464 4767 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442476 4767 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442488 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442499 4767 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442512 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442526 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442538 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442647 4767 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442661 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442675 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442688 4767 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442700 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442712 4767 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442724 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442736 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442749 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442764 4767 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442778 4767 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442791 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442805 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442817 4767 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442830 4767 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442842 4767 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442854 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442863 4767 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442847 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442872 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443038 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443054 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443064 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443075 4767 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443088 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443098 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443108 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443117 4767 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443128 4767 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443138 4767 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443147 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443156 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443166 4767 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443179 4767 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443189 4767 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443210 4767 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443220 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443229 4767 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443239 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443249 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443258 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443268 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443278 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443287 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443301 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443311 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443320 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443329 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443338 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443347 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443357 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443366 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443374 4767 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443383 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443392 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443402 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443412 4767 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443421 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443430 4767 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443439 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443449 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443458 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.442621 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443482 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443656 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.443835 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.444013 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.444030 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.444123 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.444477 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.444761 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.444802 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.445022 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.445086 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.445366 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.445541 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.445599 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:24.945582177 +0000 UTC m=+96.358898224 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.445789 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.445970 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.445977 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.445800 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.449108 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:24.949089168 +0000 UTC m=+96.362405215 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.447382 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.447146 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.449179 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.447394 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.447607 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.447650 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.447750 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.447944 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.448659 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.448685 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.449878 4767 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.450435 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.450535 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.450886 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.452302 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.455657 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.455908 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.456089 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.456311 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.456515 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.458763 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.459229 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.459257 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.459584 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.459600 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.459610 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.459667 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:24.959646491 +0000 UTC m=+96.372962538 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.459674 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.460381 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.460628 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.460893 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.460903 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.461142 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.461355 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.461602 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.461896 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.463062 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.463587 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.463835 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.463631 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.464052 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.463860 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.464617 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.464898 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.465062 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.465076 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.465119 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.465169 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:24.965151519 +0000 UTC m=+96.378467566 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.465423 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.465551 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.468481 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.468486 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.468914 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.429590 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.469571 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.469629 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.469731 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.470020 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.470862 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.470867 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.471014 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.471054 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.471138 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.471409 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.471621 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.471653 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.471664 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.471679 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.471690 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:24Z","lastTransitionTime":"2026-03-17T15:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.472362 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.472537 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.472870 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.473040 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.473094 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.473117 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.473283 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.473305 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.473515 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.474066 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.474242 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.474365 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.474378 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.474419 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.474722 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.474869 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.475122 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.475387 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.475409 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.475629 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.475979 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.476015 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.476512 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.477614 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.477761 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.477950 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.478168 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.478425 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.478730 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.478739 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.480137 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.481211 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.484209 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.489363 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.492851 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.493846 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.495376 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.500126 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.506855 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.513450 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.526117 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.537603 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544620 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544688 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544732 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544745 4767 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544757 4767 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544768 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544779 4767 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544791 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544801 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544814 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544805 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544827 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544939 4767 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544954 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544967 4767 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544982 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.544993 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545005 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545018 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545029 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545040 4767 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545051 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545062 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545072 4767 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545086 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545099 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545110 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545122 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545133 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545145 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545155 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545166 4767 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545195 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545211 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545224 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545236 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545247 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545259 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545273 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545265 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545286 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545446 4767 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545480 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545545 4767 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545573 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545648 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545678 4767 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545743 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545772 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545836 4767 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545857 4767 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545875 4767 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545925 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545944 4767 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.545961 4767 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546011 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546030 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546049 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546097 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546118 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546139 4767 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546214 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546231 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546251 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546304 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546328 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546394 4767 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546424 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546444 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546496 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546516 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546538 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546605 4767 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546633 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546705 4767 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546731 4767 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546797 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546816 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546867 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546887 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546904 4767 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546974 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.546892 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547001 4767 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547114 4767 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547203 4767 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547232 4767 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547256 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547326 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547354 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547420 4767 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547446 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547511 4767 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547536 4767 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547646 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547718 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547744 4767 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547802 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547820 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547840 4767 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547891 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547908 4767 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547926 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.547944 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.558074 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.567407 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.574312 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.574350 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.574362 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.574379 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.574390 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:24Z","lastTransitionTime":"2026-03-17T15:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.676820 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.676877 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.676894 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.676985 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.677005 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:24Z","lastTransitionTime":"2026-03-17T15:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.679135 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.689356 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.699283 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 15:38:24 crc kubenswrapper[4767]: W0317 15:38:24.703327 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-1c8cb6a3e6f0b14072dae78216aa5cf850fa09a31d64bffc822f8287dcf0bda9 WatchSource:0}: Error finding container 1c8cb6a3e6f0b14072dae78216aa5cf850fa09a31d64bffc822f8287dcf0bda9: Status 404 returned error can't find the container with id 1c8cb6a3e6f0b14072dae78216aa5cf850fa09a31d64bffc822f8287dcf0bda9 Mar 17 15:38:24 crc kubenswrapper[4767]: W0317 15:38:24.721774 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-370c8d238b28eb3c643e80a5fc4ae6e77254bc9b12a0ab67c84241cf079a428a WatchSource:0}: Error finding container 370c8d238b28eb3c643e80a5fc4ae6e77254bc9b12a0ab67c84241cf079a428a: Status 404 returned error can't find the container with id 370c8d238b28eb3c643e80a5fc4ae6e77254bc9b12a0ab67c84241cf079a428a Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.725797 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1c8cb6a3e6f0b14072dae78216aa5cf850fa09a31d64bffc822f8287dcf0bda9"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.727476 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a38a50965de1c109600897d1e2f705354176be55bfdcbc4cb7fe70cd87fe0df3"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.728497 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"370c8d238b28eb3c643e80a5fc4ae6e77254bc9b12a0ab67c84241cf079a428a"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.780880 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.780921 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.780934 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.780951 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.780962 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:24Z","lastTransitionTime":"2026-03-17T15:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.883324 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.883440 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.883524 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.883697 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.883770 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:24Z","lastTransitionTime":"2026-03-17T15:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.951904 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.951977 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.951998 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.952109 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.952150 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:25.95213807 +0000 UTC m=+97.365454117 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.952208 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:38:25.952202472 +0000 UTC m=+97.365518519 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.952235 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: E0317 15:38:24.952255 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:25.952248473 +0000 UTC m=+97.365564520 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.986647 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.986687 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.986698 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.986712 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:24 crc kubenswrapper[4767]: I0317 15:38:24.986723 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:24Z","lastTransitionTime":"2026-03-17T15:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.053468 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.053746 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.053654 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.054010 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.054130 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.054318 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:26.054294967 +0000 UTC m=+97.467611024 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.053828 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.054598 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.056266 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.056373 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:26.056352016 +0000 UTC m=+97.469668123 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.088542 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.088578 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.088590 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.088605 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.088617 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:25Z","lastTransitionTime":"2026-03-17T15:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.191600 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.192157 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.192318 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.192456 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.192583 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:25Z","lastTransitionTime":"2026-03-17T15:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.295734 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.295776 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.295784 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.295796 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.295805 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:25Z","lastTransitionTime":"2026-03-17T15:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.358872 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.359490 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.360910 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.361608 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.362556 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.363059 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.363674 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.364634 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.365344 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.366255 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.366742 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.367731 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.368227 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.368703 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.369584 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.370079 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.371041 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.371432 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.371999 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.373066 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.373568 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.374672 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.375123 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.376267 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.376760 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.377442 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.378535 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.379033 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.379956 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.380479 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.381492 4767 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.381640 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.383423 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.384462 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.384964 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.386908 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.387926 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.389013 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.389755 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.390829 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.391483 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.392732 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.393533 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.394544 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.394991 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.396168 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.396706 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.397930 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.398415 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.398468 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.398479 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.398496 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.398508 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:25Z","lastTransitionTime":"2026-03-17T15:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.398440 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.399503 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.399952 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.400969 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.401694 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.402378 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.501156 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.501215 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.501227 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.501243 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.501254 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:25Z","lastTransitionTime":"2026-03-17T15:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.604412 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.604732 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.604806 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.604882 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.604949 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:25Z","lastTransitionTime":"2026-03-17T15:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.708837 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.708897 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.708921 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.708947 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.708962 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:25Z","lastTransitionTime":"2026-03-17T15:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.734619 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.734694 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.737089 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.751120 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.765944 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.777649 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.788203 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.796364 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.805668 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.811608 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.811653 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.811666 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.811683 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.811695 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:25Z","lastTransitionTime":"2026-03-17T15:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.818951 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.829282 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.840658 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.852566 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.863644 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.875072 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.914660 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.914710 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.914719 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.914734 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.914745 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:25Z","lastTransitionTime":"2026-03-17T15:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.964467 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.964520 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:38:27.964497284 +0000 UTC m=+99.377813331 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.964591 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:25 crc kubenswrapper[4767]: I0317 15:38:25.964621 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.964714 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.964727 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.964770 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:27.964757952 +0000 UTC m=+99.378073999 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:25 crc kubenswrapper[4767]: E0317 15:38:25.964791 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:27.964782253 +0000 UTC m=+99.378098300 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.016894 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.016927 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.016936 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.016951 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.016960 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:26Z","lastTransitionTime":"2026-03-17T15:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.065817 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.065888 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:26 crc kubenswrapper[4767]: E0317 15:38:26.065963 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:26 crc kubenswrapper[4767]: E0317 15:38:26.065985 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:26 crc kubenswrapper[4767]: E0317 15:38:26.065986 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:26 crc kubenswrapper[4767]: E0317 15:38:26.065998 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:26 crc kubenswrapper[4767]: E0317 15:38:26.066004 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:26 crc kubenswrapper[4767]: E0317 15:38:26.066015 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:26 crc kubenswrapper[4767]: E0317 15:38:26.066049 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:28.066035754 +0000 UTC m=+99.479351801 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:26 crc kubenswrapper[4767]: E0317 15:38:26.066074 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:28.066058224 +0000 UTC m=+99.479374271 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.118992 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.119034 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.119083 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.119110 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.119157 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:26Z","lastTransitionTime":"2026-03-17T15:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.223266 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.223326 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.223343 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.223366 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.223386 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:26Z","lastTransitionTime":"2026-03-17T15:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.325452 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.325496 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.325512 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.325536 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.325551 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:26Z","lastTransitionTime":"2026-03-17T15:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.353436 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.353478 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.353447 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:26 crc kubenswrapper[4767]: E0317 15:38:26.353574 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:26 crc kubenswrapper[4767]: E0317 15:38:26.353644 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:26 crc kubenswrapper[4767]: E0317 15:38:26.353772 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.427449 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.427498 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.427515 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.427537 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.427553 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:26Z","lastTransitionTime":"2026-03-17T15:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.529969 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.530009 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.530034 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.530050 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.530087 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:26Z","lastTransitionTime":"2026-03-17T15:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.632461 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.632532 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.632543 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.632565 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.632576 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:26Z","lastTransitionTime":"2026-03-17T15:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.735392 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.735460 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.735474 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.735494 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.735527 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:26Z","lastTransitionTime":"2026-03-17T15:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.838086 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.838129 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.838146 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.838188 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.838200 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:26Z","lastTransitionTime":"2026-03-17T15:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.940232 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.940267 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.940276 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.940288 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:26 crc kubenswrapper[4767]: I0317 15:38:26.940298 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:26Z","lastTransitionTime":"2026-03-17T15:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.042489 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.042536 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.042552 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.042575 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.042592 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:27Z","lastTransitionTime":"2026-03-17T15:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.145041 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.145070 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.145077 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.145089 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.145099 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:27Z","lastTransitionTime":"2026-03-17T15:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.247604 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.247662 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.247674 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.247689 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.247720 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:27Z","lastTransitionTime":"2026-03-17T15:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.350454 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.350490 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.350498 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.350512 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.350521 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:27Z","lastTransitionTime":"2026-03-17T15:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.453561 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.453591 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.453599 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.453629 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.453638 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:27Z","lastTransitionTime":"2026-03-17T15:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.556608 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.556692 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.556708 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.556762 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.556778 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:27Z","lastTransitionTime":"2026-03-17T15:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.659082 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.659112 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.659120 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.659132 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.659140 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:27Z","lastTransitionTime":"2026-03-17T15:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.743553 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.763340 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.763803 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.764124 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.764968 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.765484 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.766002 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:27Z","lastTransitionTime":"2026-03-17T15:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.782856 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.799648 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.821314 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.846669 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.861468 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.869524 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.869576 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.869594 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.869613 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.869627 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:27Z","lastTransitionTime":"2026-03-17T15:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.972243 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.972286 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.972296 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.972311 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.972320 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:27Z","lastTransitionTime":"2026-03-17T15:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.983073 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:27 crc kubenswrapper[4767]: E0317 15:38:27.983264 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:38:31.983243225 +0000 UTC m=+103.396559282 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.983415 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:27 crc kubenswrapper[4767]: E0317 15:38:27.983534 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:27 crc kubenswrapper[4767]: E0317 15:38:27.983572 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:31.983564744 +0000 UTC m=+103.396880791 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:27 crc kubenswrapper[4767]: I0317 15:38:27.983531 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:27 crc kubenswrapper[4767]: E0317 15:38:27.983783 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:27 crc kubenswrapper[4767]: E0317 15:38:27.983916 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:31.983899254 +0000 UTC m=+103.397215301 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.075087 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.075418 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.075517 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.075607 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.075685 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:28Z","lastTransitionTime":"2026-03-17T15:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.084849 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.085042 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:28 crc kubenswrapper[4767]: E0317 15:38:28.085269 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:28 crc kubenswrapper[4767]: E0317 15:38:28.085361 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:28 crc kubenswrapper[4767]: E0317 15:38:28.085458 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:28 crc kubenswrapper[4767]: E0317 15:38:28.085591 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:32.085573747 +0000 UTC m=+103.498889794 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:28 crc kubenswrapper[4767]: E0317 15:38:28.085373 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:28 crc kubenswrapper[4767]: E0317 15:38:28.085750 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:28 crc kubenswrapper[4767]: E0317 15:38:28.085822 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:28 crc kubenswrapper[4767]: E0317 15:38:28.085918 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:32.085907567 +0000 UTC m=+103.499223624 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.177869 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.177943 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.177962 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.177987 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.178004 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:28Z","lastTransitionTime":"2026-03-17T15:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.281149 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.281217 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.281226 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.281240 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.281250 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:28Z","lastTransitionTime":"2026-03-17T15:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.354311 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.354331 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:28 crc kubenswrapper[4767]: E0317 15:38:28.354538 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.354333 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:28 crc kubenswrapper[4767]: E0317 15:38:28.354649 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:28 crc kubenswrapper[4767]: E0317 15:38:28.354832 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.384067 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.384111 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.384127 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.384149 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.384197 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:28Z","lastTransitionTime":"2026-03-17T15:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.487020 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.487851 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.487981 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.488081 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.488171 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:28Z","lastTransitionTime":"2026-03-17T15:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.591626 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.591666 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.591682 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.591704 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.591720 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:28Z","lastTransitionTime":"2026-03-17T15:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.694357 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.694396 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.694404 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.694418 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.694428 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:28Z","lastTransitionTime":"2026-03-17T15:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.796641 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.796704 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.796714 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.796732 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.796750 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:28Z","lastTransitionTime":"2026-03-17T15:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.899031 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.899076 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.899089 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.899131 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:28 crc kubenswrapper[4767]: I0317 15:38:28.899142 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:28Z","lastTransitionTime":"2026-03-17T15:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.002237 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.002595 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.002812 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.003023 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.003285 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:29Z","lastTransitionTime":"2026-03-17T15:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.105866 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.105913 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.105926 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.105942 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.105957 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:29Z","lastTransitionTime":"2026-03-17T15:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.209476 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.209552 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.209574 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.209604 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.209626 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:29Z","lastTransitionTime":"2026-03-17T15:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.311584 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.311625 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.311635 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.311649 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.311659 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:29Z","lastTransitionTime":"2026-03-17T15:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.368980 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.379774 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.392585 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.406021 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.413378 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.413403 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.413413 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.413426 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.413434 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:29Z","lastTransitionTime":"2026-03-17T15:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.421874 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.433761 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.515619 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.515650 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.515658 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.515670 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.515679 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:29Z","lastTransitionTime":"2026-03-17T15:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.618035 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.618082 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.618094 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.618110 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.618122 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:29Z","lastTransitionTime":"2026-03-17T15:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.721151 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.721181 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.721204 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.721216 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.721224 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:29Z","lastTransitionTime":"2026-03-17T15:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.824489 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.824523 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.824533 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.824548 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.824558 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:29Z","lastTransitionTime":"2026-03-17T15:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.927282 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.927309 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.927317 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.927328 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:29 crc kubenswrapper[4767]: I0317 15:38:29.927336 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:29Z","lastTransitionTime":"2026-03-17T15:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.029334 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.029367 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.029376 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.029391 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.029402 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:30Z","lastTransitionTime":"2026-03-17T15:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.132142 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.132252 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.132266 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.132283 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.132617 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:30Z","lastTransitionTime":"2026-03-17T15:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.235316 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.235350 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.235361 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.235376 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.235408 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:30Z","lastTransitionTime":"2026-03-17T15:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.338040 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.338113 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.338141 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.338176 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.338248 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:30Z","lastTransitionTime":"2026-03-17T15:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.353968 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:30 crc kubenswrapper[4767]: E0317 15:38:30.354207 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.354704 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:30 crc kubenswrapper[4767]: E0317 15:38:30.354806 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.354887 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:30 crc kubenswrapper[4767]: E0317 15:38:30.355005 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.374690 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.440603 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.440637 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.440648 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.440663 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.440675 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:30Z","lastTransitionTime":"2026-03-17T15:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.543549 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.543592 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.543609 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.543630 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.543647 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:30Z","lastTransitionTime":"2026-03-17T15:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.646265 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.646310 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.646319 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.646333 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.646342 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:30Z","lastTransitionTime":"2026-03-17T15:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.749740 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.750004 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.750109 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.750267 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.750371 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:30Z","lastTransitionTime":"2026-03-17T15:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.852956 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.853199 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.853384 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.853529 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.853667 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:30Z","lastTransitionTime":"2026-03-17T15:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.957404 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.957697 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.957845 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.957995 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:30 crc kubenswrapper[4767]: I0317 15:38:30.958158 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:30Z","lastTransitionTime":"2026-03-17T15:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.060800 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.060840 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.060850 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.060867 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.060878 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:31Z","lastTransitionTime":"2026-03-17T15:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.164211 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.164291 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.164315 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.164346 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.164370 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:31Z","lastTransitionTime":"2026-03-17T15:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.267628 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.267671 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.267680 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.267696 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.267705 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:31Z","lastTransitionTime":"2026-03-17T15:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.370795 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.370888 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.370914 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.370944 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.370968 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:31Z","lastTransitionTime":"2026-03-17T15:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.372618 4767 scope.go:117] "RemoveContainer" containerID="8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.372907 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.473689 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.474042 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.474057 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.474099 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.474112 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:31Z","lastTransitionTime":"2026-03-17T15:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.577011 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.577049 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.577062 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.577077 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.577089 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:31Z","lastTransitionTime":"2026-03-17T15:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.679214 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.679267 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.679282 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.679303 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.679318 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:31Z","lastTransitionTime":"2026-03-17T15:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.757347 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.759488 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2"} Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.759746 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.774305 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:31Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.781646 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.781691 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.781705 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.781723 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.781738 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:31Z","lastTransitionTime":"2026-03-17T15:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.795228 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:31Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.809255 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:31Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.830051 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:31Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.845358 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:31Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.866388 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:31Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.879917 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:31Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.884023 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.884114 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.884198 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.884275 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.884338 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:31Z","lastTransitionTime":"2026-03-17T15:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.893424 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:31Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.987699 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.987759 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.987793 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.987821 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:31 crc kubenswrapper[4767]: I0317 15:38:31.987860 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:31Z","lastTransitionTime":"2026-03-17T15:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.021246 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.021396 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.021448 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.021610 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.021690 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:40.021657539 +0000 UTC m=+111.434973606 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.022267 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:38:40.022252166 +0000 UTC m=+111.435568223 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.022326 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.022360 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:40.022349849 +0000 UTC m=+111.435665906 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.090078 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.090106 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.090114 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.090126 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.090134 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:32Z","lastTransitionTime":"2026-03-17T15:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.122225 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.122445 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.122496 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.122509 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.122571 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:40.122548209 +0000 UTC m=+111.535864336 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.122476 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.122931 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.123058 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.123167 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.123395 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:40.123371973 +0000 UTC m=+111.536688060 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.192694 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.192964 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.193036 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.193096 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.193150 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:32Z","lastTransitionTime":"2026-03-17T15:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.296856 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.297107 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.297209 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.297393 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.297481 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:32Z","lastTransitionTime":"2026-03-17T15:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.353933 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.354224 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.354386 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.354422 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.354517 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:32 crc kubenswrapper[4767]: E0317 15:38:32.354586 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.400070 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.400111 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.400120 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.400133 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.400144 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:32Z","lastTransitionTime":"2026-03-17T15:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.502842 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.502885 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.502893 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.502908 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.502918 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:32Z","lastTransitionTime":"2026-03-17T15:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.605551 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.606150 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.606287 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.606388 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.606502 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:32Z","lastTransitionTime":"2026-03-17T15:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.710159 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.710208 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.710219 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.710233 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.710244 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:32Z","lastTransitionTime":"2026-03-17T15:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.813049 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.813287 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.813349 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.813415 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.813472 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:32Z","lastTransitionTime":"2026-03-17T15:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.915864 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.915914 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.915925 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.915942 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:32 crc kubenswrapper[4767]: I0317 15:38:32.915953 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:32Z","lastTransitionTime":"2026-03-17T15:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.019338 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.019407 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.019425 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.019451 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.019470 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.123209 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.123273 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.123290 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.123316 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.123336 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.225830 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.225860 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.225867 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.225879 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.225888 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.328569 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.328607 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.328619 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.328634 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.328648 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.438006 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.438081 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.438106 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.438142 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.438171 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.477060 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-l7694"] Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.477424 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-l7694" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.480847 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.481390 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.481837 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.502194 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.517851 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.531149 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.540071 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.540138 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.540162 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.540230 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.540255 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.548008 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.559875 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.580546 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.593031 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.607851 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.621110 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.633689 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fn87\" (UniqueName: \"kubernetes.io/projected/6c136104-991b-4c97-ad84-80b82b617b0b-kube-api-access-8fn87\") pod \"node-resolver-l7694\" (UID: \"6c136104-991b-4c97-ad84-80b82b617b0b\") " pod="openshift-dns/node-resolver-l7694" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.633761 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6c136104-991b-4c97-ad84-80b82b617b0b-hosts-file\") pod \"node-resolver-l7694\" (UID: \"6c136104-991b-4c97-ad84-80b82b617b0b\") " pod="openshift-dns/node-resolver-l7694" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.643469 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.643510 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.643519 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.643536 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.643548 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.649824 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.649862 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.649871 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.649883 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.649894 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: E0317 15:38:33.664300 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.669120 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.669157 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.669174 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.669208 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.669222 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: E0317 15:38:33.684104 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.689414 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.689479 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.689496 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.689519 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.689536 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: E0317 15:38:33.704626 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.709363 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.709426 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.709440 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.709463 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.709481 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: E0317 15:38:33.722462 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.727127 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.727192 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.727206 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.727225 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.727236 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.734337 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fn87\" (UniqueName: \"kubernetes.io/projected/6c136104-991b-4c97-ad84-80b82b617b0b-kube-api-access-8fn87\") pod \"node-resolver-l7694\" (UID: \"6c136104-991b-4c97-ad84-80b82b617b0b\") " pod="openshift-dns/node-resolver-l7694" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.734435 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6c136104-991b-4c97-ad84-80b82b617b0b-hosts-file\") pod \"node-resolver-l7694\" (UID: \"6c136104-991b-4c97-ad84-80b82b617b0b\") " pod="openshift-dns/node-resolver-l7694" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.734567 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6c136104-991b-4c97-ad84-80b82b617b0b-hosts-file\") pod \"node-resolver-l7694\" (UID: \"6c136104-991b-4c97-ad84-80b82b617b0b\") " pod="openshift-dns/node-resolver-l7694" Mar 17 15:38:33 crc kubenswrapper[4767]: E0317 15:38:33.742504 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: E0317 15:38:33.742662 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.746353 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.746402 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.746418 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.746435 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.746447 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.757270 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fn87\" (UniqueName: \"kubernetes.io/projected/6c136104-991b-4c97-ad84-80b82b617b0b-kube-api-access-8fn87\") pod \"node-resolver-l7694\" (UID: \"6c136104-991b-4c97-ad84-80b82b617b0b\") " pod="openshift-dns/node-resolver-l7694" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.790161 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-l7694" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.837712 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-nqc7l"] Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.838205 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-v6gnj"] Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.838342 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-lhrtr"] Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.838391 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.838598 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.838788 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.841566 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.842637 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.842720 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.842656 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.842896 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.843251 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.843476 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.843837 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.844343 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.844482 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.844578 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.844698 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.850125 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.850153 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.850161 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.850188 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.850198 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.864212 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.879564 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.893193 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.901711 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.913300 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.923843 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.937331 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938124 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-os-release\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938347 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/811a8833-5459-4af6-a1c3-179040eb3c01-cni-binary-copy\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938409 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-system-cni-dir\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938443 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-multus-conf-dir\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938492 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d680dc75-619e-41ae-90cb-2f58898a33ac-proxy-tls\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938557 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-os-release\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938597 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-var-lib-cni-multus\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938641 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69rwb\" (UniqueName: \"kubernetes.io/projected/811a8833-5459-4af6-a1c3-179040eb3c01-kube-api-access-69rwb\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938689 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/811a8833-5459-4af6-a1c3-179040eb3c01-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938781 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wp98\" (UniqueName: \"kubernetes.io/projected/f36218dd-ddd3-4269-af32-7fe03113b342-kube-api-access-7wp98\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938868 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-etc-kubernetes\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938892 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f36218dd-ddd3-4269-af32-7fe03113b342-cni-binary-copy\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938908 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-var-lib-kubelet\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.938927 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d680dc75-619e-41ae-90cb-2f58898a33ac-mcd-auth-proxy-config\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939021 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-cnibin\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939095 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-run-multus-certs\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939115 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939185 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-system-cni-dir\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939203 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-var-lib-cni-bin\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939304 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d680dc75-619e-41ae-90cb-2f58898a33ac-rootfs\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939355 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-cnibin\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939400 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-run-k8s-cni-cncf-io\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939431 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-hostroot\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939477 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-multus-cni-dir\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939503 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-multus-socket-dir-parent\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939523 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-run-netns\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939556 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f36218dd-ddd3-4269-af32-7fe03113b342-multus-daemon-config\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.939581 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r97fm\" (UniqueName: \"kubernetes.io/projected/d680dc75-619e-41ae-90cb-2f58898a33ac-kube-api-access-r97fm\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.950630 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.952513 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.952563 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.952586 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.952637 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.952661 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:33Z","lastTransitionTime":"2026-03-17T15:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.963720 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:33 crc kubenswrapper[4767]: I0317 15:38:33.980917 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.001566 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:33Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.014150 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.026428 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.039693 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040260 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/811a8833-5459-4af6-a1c3-179040eb3c01-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040305 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69rwb\" (UniqueName: \"kubernetes.io/projected/811a8833-5459-4af6-a1c3-179040eb3c01-kube-api-access-69rwb\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040329 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-etc-kubernetes\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040353 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wp98\" (UniqueName: \"kubernetes.io/projected/f36218dd-ddd3-4269-af32-7fe03113b342-kube-api-access-7wp98\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040374 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f36218dd-ddd3-4269-af32-7fe03113b342-cni-binary-copy\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040393 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-cnibin\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040414 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-var-lib-kubelet\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040435 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d680dc75-619e-41ae-90cb-2f58898a33ac-mcd-auth-proxy-config\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040471 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-etc-kubernetes\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040483 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-cnibin\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040529 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-var-lib-kubelet\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040561 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-run-multus-certs\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040750 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040774 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-system-cni-dir\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040790 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-var-lib-cni-bin\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040797 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-run-multus-certs\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040809 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d680dc75-619e-41ae-90cb-2f58898a33ac-rootfs\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040828 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d680dc75-619e-41ae-90cb-2f58898a33ac-rootfs\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040907 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-var-lib-cni-bin\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041010 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-cnibin\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041031 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-system-cni-dir\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041240 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041394 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f36218dd-ddd3-4269-af32-7fe03113b342-cni-binary-copy\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041485 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d680dc75-619e-41ae-90cb-2f58898a33ac-mcd-auth-proxy-config\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.040849 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-cnibin\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041566 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-run-k8s-cni-cncf-io\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041592 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-hostroot\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041618 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r97fm\" (UniqueName: \"kubernetes.io/projected/d680dc75-619e-41ae-90cb-2f58898a33ac-kube-api-access-r97fm\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041638 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-multus-cni-dir\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041660 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-multus-socket-dir-parent\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041680 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-run-netns\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041678 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/811a8833-5459-4af6-a1c3-179040eb3c01-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041691 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-hostroot\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041701 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f36218dd-ddd3-4269-af32-7fe03113b342-multus-daemon-config\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041733 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-run-k8s-cni-cncf-io\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041742 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-multus-socket-dir-parent\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041752 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-multus-cni-dir\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041770 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-run-netns\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041850 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-os-release\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041882 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/811a8833-5459-4af6-a1c3-179040eb3c01-cni-binary-copy\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041909 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d680dc75-619e-41ae-90cb-2f58898a33ac-proxy-tls\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041929 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-system-cni-dir\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041950 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-multus-conf-dir\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.041984 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-var-lib-cni-multus\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.042018 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-os-release\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.042025 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-os-release\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.042070 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-host-var-lib-cni-multus\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.042076 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/811a8833-5459-4af6-a1c3-179040eb3c01-system-cni-dir\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.042104 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-os-release\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.042107 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f36218dd-ddd3-4269-af32-7fe03113b342-multus-conf-dir\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.042294 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f36218dd-ddd3-4269-af32-7fe03113b342-multus-daemon-config\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.042551 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/811a8833-5459-4af6-a1c3-179040eb3c01-cni-binary-copy\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.045544 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d680dc75-619e-41ae-90cb-2f58898a33ac-proxy-tls\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.054827 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.054856 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.054865 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.054878 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.054886 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:34Z","lastTransitionTime":"2026-03-17T15:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.056060 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.057040 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69rwb\" (UniqueName: \"kubernetes.io/projected/811a8833-5459-4af6-a1c3-179040eb3c01-kube-api-access-69rwb\") pod \"multus-additional-cni-plugins-nqc7l\" (UID: \"811a8833-5459-4af6-a1c3-179040eb3c01\") " pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.060555 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r97fm\" (UniqueName: \"kubernetes.io/projected/d680dc75-619e-41ae-90cb-2f58898a33ac-kube-api-access-r97fm\") pod \"machine-config-daemon-lhrtr\" (UID: \"d680dc75-619e-41ae-90cb-2f58898a33ac\") " pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.064960 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wp98\" (UniqueName: \"kubernetes.io/projected/f36218dd-ddd3-4269-af32-7fe03113b342-kube-api-access-7wp98\") pod \"multus-v6gnj\" (UID: \"f36218dd-ddd3-4269-af32-7fe03113b342\") " pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.075682 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.084322 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.096407 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.111570 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.125404 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.137112 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.149300 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.155402 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.160192 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.160236 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.160245 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.160260 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.160269 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:34Z","lastTransitionTime":"2026-03-17T15:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.165373 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.172282 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-v6gnj" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.210046 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c4n7r"] Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.211070 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.215090 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.215337 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.215474 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.215585 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.215696 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.215845 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.215988 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.236354 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: W0317 15:38:34.237151 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf36218dd_ddd3_4269_af32_7fe03113b342.slice/crio-39a9ed2cc0bb36488dca6d33a474ecfdb8066bc300338fdbd9bb8133c0a8b85d WatchSource:0}: Error finding container 39a9ed2cc0bb36488dca6d33a474ecfdb8066bc300338fdbd9bb8133c0a8b85d: Status 404 returned error can't find the container with id 39a9ed2cc0bb36488dca6d33a474ecfdb8066bc300338fdbd9bb8133c0a8b85d Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.250766 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.262829 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.262873 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.262885 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.262901 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.262914 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:34Z","lastTransitionTime":"2026-03-17T15:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.265758 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.275299 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.287631 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.302513 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.317687 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.329616 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.340102 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345370 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-etc-openvswitch\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345414 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-ovn-kubernetes\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345442 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-config\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345480 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345505 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-openvswitch\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345526 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-var-lib-openvswitch\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345548 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-netd\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345568 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-ovn\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345587 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-systemd-units\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345610 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-netns\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345644 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-log-socket\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345666 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpdb2\" (UniqueName: \"kubernetes.io/projected/93753cb4-c3fa-424e-84a2-0a7db20eecfa-kube-api-access-zpdb2\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345730 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-node-log\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345750 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-systemd\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345770 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-script-lib\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345804 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-slash\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345825 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-bin\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345847 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-env-overrides\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345912 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovn-node-metrics-cert\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.345936 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-kubelet\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.353330 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.353389 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:34 crc kubenswrapper[4767]: E0317 15:38:34.353433 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.353442 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:34 crc kubenswrapper[4767]: E0317 15:38:34.353531 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:34 crc kubenswrapper[4767]: E0317 15:38:34.353623 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.362123 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.364866 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.364895 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.364903 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.364916 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.364925 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:34Z","lastTransitionTime":"2026-03-17T15:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.383765 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.394353 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.405732 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447202 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-netd\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447265 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-ovn\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447297 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-log-socket\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447327 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpdb2\" (UniqueName: \"kubernetes.io/projected/93753cb4-c3fa-424e-84a2-0a7db20eecfa-kube-api-access-zpdb2\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447358 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-systemd-units\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447385 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-netns\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447430 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-node-log\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447458 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-systemd\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447484 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-script-lib\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447526 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-slash\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447556 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-bin\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447583 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-env-overrides\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447614 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovn-node-metrics-cert\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447648 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-kubelet\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447693 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-etc-openvswitch\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447723 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-ovn-kubernetes\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447769 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-config\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447813 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447847 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-openvswitch\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447877 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-var-lib-openvswitch\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.447969 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-var-lib-openvswitch\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.448032 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-netd\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.448075 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-ovn\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.448114 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-log-socket\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.449505 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-env-overrides\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.449581 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-systemd-units\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.449613 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-netns\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.449643 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-node-log\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.449671 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-systemd\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.450276 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-script-lib\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.450319 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-slash\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.450347 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-bin\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.450377 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-ovn-kubernetes\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.450400 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-kubelet\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.450422 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-etc-openvswitch\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.450447 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.450942 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-config\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.450979 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-openvswitch\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.454435 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovn-node-metrics-cert\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.467575 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.467606 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.467616 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.467629 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.467638 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:34Z","lastTransitionTime":"2026-03-17T15:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.471301 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpdb2\" (UniqueName: \"kubernetes.io/projected/93753cb4-c3fa-424e-84a2-0a7db20eecfa-kube-api-access-zpdb2\") pod \"ovnkube-node-c4n7r\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.528508 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.569547 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.569577 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.569585 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.569597 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.569606 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:34Z","lastTransitionTime":"2026-03-17T15:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.671439 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.671727 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.671735 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.671751 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.671762 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:34Z","lastTransitionTime":"2026-03-17T15:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.772434 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" event={"ID":"811a8833-5459-4af6-a1c3-179040eb3c01","Type":"ContainerStarted","Data":"d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.772474 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" event={"ID":"811a8833-5459-4af6-a1c3-179040eb3c01","Type":"ContainerStarted","Data":"34dbd65373ddba9fa4e087a5152a9a21a4c4e0e7ccff39691051bf86f68027fd"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.773974 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.774020 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.774035 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.774058 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.774073 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:34Z","lastTransitionTime":"2026-03-17T15:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.776292 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-l7694" event={"ID":"6c136104-991b-4c97-ad84-80b82b617b0b","Type":"ContainerStarted","Data":"3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.776329 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-l7694" event={"ID":"6c136104-991b-4c97-ad84-80b82b617b0b","Type":"ContainerStarted","Data":"110b9d12daa6211ec11f99cc32cd6b8a530c4dbc159ab246338aab558bdce4a2"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.777505 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-v6gnj" event={"ID":"f36218dd-ddd3-4269-af32-7fe03113b342","Type":"ContainerStarted","Data":"6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.777535 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-v6gnj" event={"ID":"f36218dd-ddd3-4269-af32-7fe03113b342","Type":"ContainerStarted","Data":"39a9ed2cc0bb36488dca6d33a474ecfdb8066bc300338fdbd9bb8133c0a8b85d"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.779063 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff" exitCode=0 Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.779124 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.779150 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"9d31966e2e47c089ff7da20e99a504f1676eac18a67ebc2a715662046186f2c6"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.783348 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.783404 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.783424 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"0449dde59a6fdb06203560a4b7419bbd0b62c68f25697118221a8bf5b6f837a6"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.793085 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.812739 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.828255 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.848353 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.862615 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.875808 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.875838 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.875847 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.875861 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.875869 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:34Z","lastTransitionTime":"2026-03-17T15:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.885802 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.897370 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.908680 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.922516 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.936800 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.950048 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.961021 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.975842 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.978699 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.978734 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.978745 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.978764 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:34 crc kubenswrapper[4767]: I0317 15:38:34.978777 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:34Z","lastTransitionTime":"2026-03-17T15:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:34.999209 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:34Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.042414 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.061307 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.083465 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.083509 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.083528 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.083554 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.083574 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:35Z","lastTransitionTime":"2026-03-17T15:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.088484 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.110163 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.125023 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.133580 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.144992 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.153816 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.189403 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.189450 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.189459 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.189472 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.189494 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:35Z","lastTransitionTime":"2026-03-17T15:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.189895 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.204313 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.218435 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.233801 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.301938 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.301971 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.301979 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.301996 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.302007 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:35Z","lastTransitionTime":"2026-03-17T15:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.404470 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.404683 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.404760 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.404832 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.404904 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:35Z","lastTransitionTime":"2026-03-17T15:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.507586 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.507617 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.507625 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.507637 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.507646 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:35Z","lastTransitionTime":"2026-03-17T15:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.627341 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.627564 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.627664 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.627784 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.627978 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:35Z","lastTransitionTime":"2026-03-17T15:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.851747 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.852278 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.852387 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.852468 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.852540 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:35Z","lastTransitionTime":"2026-03-17T15:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.855592 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.855738 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.855821 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.857031 4767 generic.go:334] "Generic (PLEG): container finished" podID="811a8833-5459-4af6-a1c3-179040eb3c01" containerID="d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1" exitCode=0 Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.857132 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" event={"ID":"811a8833-5459-4af6-a1c3-179040eb3c01","Type":"ContainerDied","Data":"d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.877770 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.893021 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.953556 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.954696 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.954730 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.954738 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.954752 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.954764 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:35Z","lastTransitionTime":"2026-03-17T15:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.970376 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:35 crc kubenswrapper[4767]: I0317 15:38:35.985303 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.002535 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:35Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.015757 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:36Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.035931 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:36Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.050607 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:36Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.057529 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.057551 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.057560 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.057585 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.057599 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:36Z","lastTransitionTime":"2026-03-17T15:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.063616 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:36Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.078299 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:36Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.092476 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:36Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.116287 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:36Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.162161 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.162214 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.162224 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.162237 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.162246 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:36Z","lastTransitionTime":"2026-03-17T15:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.267990 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.268023 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.268032 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.268043 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.268052 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:36Z","lastTransitionTime":"2026-03-17T15:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.359327 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:36 crc kubenswrapper[4767]: E0317 15:38:36.359418 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.359448 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:36 crc kubenswrapper[4767]: E0317 15:38:36.359488 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.359338 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:36 crc kubenswrapper[4767]: E0317 15:38:36.359530 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.387206 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.387412 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.387420 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.387432 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.387440 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:36Z","lastTransitionTime":"2026-03-17T15:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.489522 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.489655 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.489720 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.489778 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.489830 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:36Z","lastTransitionTime":"2026-03-17T15:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.591830 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.592022 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.592099 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.592190 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.592266 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:36Z","lastTransitionTime":"2026-03-17T15:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.702290 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.702334 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.702343 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.702376 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.702395 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:36Z","lastTransitionTime":"2026-03-17T15:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.815661 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.815700 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.815709 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.815721 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.815730 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:36Z","lastTransitionTime":"2026-03-17T15:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.918704 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.918727 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.918735 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.918747 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.918755 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:36Z","lastTransitionTime":"2026-03-17T15:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.952010 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" event={"ID":"811a8833-5459-4af6-a1c3-179040eb3c01","Type":"ContainerStarted","Data":"82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.955629 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.955657 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.955666 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde"} Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.972080 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:36Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.984296 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:36Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:36 crc kubenswrapper[4767]: I0317 15:38:36.995373 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:36Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.010263 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:37Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.020795 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.020820 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.020828 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.020840 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.020857 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:37Z","lastTransitionTime":"2026-03-17T15:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.021448 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:37Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.034310 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:37Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.045584 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:37Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.059527 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:37Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.073918 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:37Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.088219 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:37Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.099850 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:37Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.109264 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:37Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.123163 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.123216 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.123226 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.123238 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.123247 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:37Z","lastTransitionTime":"2026-03-17T15:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.137450 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:37Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.226221 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.226266 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.226280 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.226297 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.226309 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:37Z","lastTransitionTime":"2026-03-17T15:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.328373 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.328413 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.328426 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.328442 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.328454 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:37Z","lastTransitionTime":"2026-03-17T15:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.444561 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.444599 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.444608 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.444622 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.444630 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:37Z","lastTransitionTime":"2026-03-17T15:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.668432 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.668477 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.668487 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.668502 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.668511 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:37Z","lastTransitionTime":"2026-03-17T15:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.770805 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.770839 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.770851 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.770867 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.770879 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:37Z","lastTransitionTime":"2026-03-17T15:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.873329 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.873360 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.873376 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.873404 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:37 crc kubenswrapper[4767]: I0317 15:38:37.873418 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:37Z","lastTransitionTime":"2026-03-17T15:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:37.975928 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:37.975968 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:37.975977 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:37.975992 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:37.976002 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:37Z","lastTransitionTime":"2026-03-17T15:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.078711 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.078961 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.078969 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.078980 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.078988 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:38Z","lastTransitionTime":"2026-03-17T15:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.181286 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.181317 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.181328 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.181343 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.181364 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:38Z","lastTransitionTime":"2026-03-17T15:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.284676 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.284732 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.284751 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.284782 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.284802 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:38Z","lastTransitionTime":"2026-03-17T15:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.353543 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.353573 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:38 crc kubenswrapper[4767]: E0317 15:38:38.353677 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.353554 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:38 crc kubenswrapper[4767]: E0317 15:38:38.353741 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:38 crc kubenswrapper[4767]: E0317 15:38:38.353818 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.387291 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.387321 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.387328 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.387340 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.387349 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:38Z","lastTransitionTime":"2026-03-17T15:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.489859 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.489897 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.489921 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.489937 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.489948 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:38Z","lastTransitionTime":"2026-03-17T15:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.592119 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.592221 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.592236 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.592258 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.592273 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:38Z","lastTransitionTime":"2026-03-17T15:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.871242 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.871288 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.871297 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.871309 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:38 crc kubenswrapper[4767]: I0317 15:38:38.871318 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:38Z","lastTransitionTime":"2026-03-17T15:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.000015 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.000246 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.000335 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.000414 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.000518 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:39Z","lastTransitionTime":"2026-03-17T15:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.086122 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.089841 4767 generic.go:334] "Generic (PLEG): container finished" podID="811a8833-5459-4af6-a1c3-179040eb3c01" containerID="82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47" exitCode=0 Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.089878 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" event={"ID":"811a8833-5459-4af6-a1c3-179040eb3c01","Type":"ContainerDied","Data":"82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.102756 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.103375 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.103496 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.103615 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.103710 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:39Z","lastTransitionTime":"2026-03-17T15:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.125158 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.143320 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.164801 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.182736 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.205648 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.252965 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.252998 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.253008 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.253022 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.253032 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:39Z","lastTransitionTime":"2026-03-17T15:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.259536 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.336867 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.351769 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.358417 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.358441 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.358449 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.358460 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.358469 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:39Z","lastTransitionTime":"2026-03-17T15:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.366462 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.383433 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.400145 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.411949 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.424775 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.437562 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.449470 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.460516 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.460579 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.460609 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.460618 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.460631 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.460647 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:39Z","lastTransitionTime":"2026-03-17T15:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.476793 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.497342 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.556003 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.569429 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.569464 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.569480 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.569495 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.569507 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:39Z","lastTransitionTime":"2026-03-17T15:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.593602 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.608236 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.621301 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.631164 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.644411 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.654200 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.666350 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:39Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.672266 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.672304 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.672314 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.672327 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.672336 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:39Z","lastTransitionTime":"2026-03-17T15:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.775113 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.775164 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.775188 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.775203 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.775215 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:39Z","lastTransitionTime":"2026-03-17T15:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.878982 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.879045 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.879059 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.879078 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.879096 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:39Z","lastTransitionTime":"2026-03-17T15:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.986682 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.986769 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.986791 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.986820 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:39 crc kubenswrapper[4767]: I0317 15:38:39.986840 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:39Z","lastTransitionTime":"2026-03-17T15:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.043562 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.043816 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.043890 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.044144 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.044247 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:56.044225505 +0000 UTC m=+127.457541552 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.044683 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.045016 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:56.044968417 +0000 UTC m=+127.458284474 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.045309 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:38:56.045294726 +0000 UTC m=+127.458610793 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.090032 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.091314 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.091350 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.091369 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.091578 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:40Z","lastTransitionTime":"2026-03-17T15:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.098851 4767 generic.go:334] "Generic (PLEG): container finished" podID="811a8833-5459-4af6-a1c3-179040eb3c01" containerID="cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1" exitCode=0 Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.098918 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" event={"ID":"811a8833-5459-4af6-a1c3-179040eb3c01","Type":"ContainerDied","Data":"cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1"} Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.129930 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.145464 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.145510 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.145650 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.145668 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.145678 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.145727 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:56.145709192 +0000 UTC m=+127.559025239 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.145785 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.145851 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.145880 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.146022 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:56.145962029 +0000 UTC m=+127.559278216 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.147123 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.164115 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.182921 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.196898 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.196964 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.196989 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.197009 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.197022 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:40Z","lastTransitionTime":"2026-03-17T15:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.197398 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.216097 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.227285 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.239926 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.252705 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.265098 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.277866 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.288087 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.299001 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.299061 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.299073 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.299097 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.299110 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:40Z","lastTransitionTime":"2026-03-17T15:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.307949 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.353704 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.353796 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.353704 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.353837 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.353916 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:40 crc kubenswrapper[4767]: E0317 15:38:40.354041 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.401603 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.401629 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.401636 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.401647 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.401666 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:40Z","lastTransitionTime":"2026-03-17T15:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.505842 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.505908 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.505925 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.505967 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.505985 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:40Z","lastTransitionTime":"2026-03-17T15:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.610938 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.611002 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.611025 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.611054 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.611075 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:40Z","lastTransitionTime":"2026-03-17T15:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.713014 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.713039 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.713050 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.713065 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.713077 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:40Z","lastTransitionTime":"2026-03-17T15:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.722725 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-gdgdx"] Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.723228 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gdgdx" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.725449 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.727554 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.727812 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.728024 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.745748 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.760740 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.780415 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.794603 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.806404 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5d82368b-ed78-4142-b1d6-9d86e0aaa810-host\") pod \"node-ca-gdgdx\" (UID: \"5d82368b-ed78-4142-b1d6-9d86e0aaa810\") " pod="openshift-image-registry/node-ca-gdgdx" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.806465 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5d82368b-ed78-4142-b1d6-9d86e0aaa810-serviceca\") pod \"node-ca-gdgdx\" (UID: \"5d82368b-ed78-4142-b1d6-9d86e0aaa810\") " pod="openshift-image-registry/node-ca-gdgdx" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.806514 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqd4h\" (UniqueName: \"kubernetes.io/projected/5d82368b-ed78-4142-b1d6-9d86e0aaa810-kube-api-access-cqd4h\") pod \"node-ca-gdgdx\" (UID: \"5d82368b-ed78-4142-b1d6-9d86e0aaa810\") " pod="openshift-image-registry/node-ca-gdgdx" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.815557 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.815623 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.815641 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.815663 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.815687 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:40Z","lastTransitionTime":"2026-03-17T15:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.813769 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.830770 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.846964 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.859906 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.871235 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.888871 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.902338 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.907289 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqd4h\" (UniqueName: \"kubernetes.io/projected/5d82368b-ed78-4142-b1d6-9d86e0aaa810-kube-api-access-cqd4h\") pod \"node-ca-gdgdx\" (UID: \"5d82368b-ed78-4142-b1d6-9d86e0aaa810\") " pod="openshift-image-registry/node-ca-gdgdx" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.907351 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5d82368b-ed78-4142-b1d6-9d86e0aaa810-host\") pod \"node-ca-gdgdx\" (UID: \"5d82368b-ed78-4142-b1d6-9d86e0aaa810\") " pod="openshift-image-registry/node-ca-gdgdx" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.907370 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5d82368b-ed78-4142-b1d6-9d86e0aaa810-serviceca\") pod \"node-ca-gdgdx\" (UID: \"5d82368b-ed78-4142-b1d6-9d86e0aaa810\") " pod="openshift-image-registry/node-ca-gdgdx" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.908130 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5d82368b-ed78-4142-b1d6-9d86e0aaa810-host\") pod \"node-ca-gdgdx\" (UID: \"5d82368b-ed78-4142-b1d6-9d86e0aaa810\") " pod="openshift-image-registry/node-ca-gdgdx" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.908311 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5d82368b-ed78-4142-b1d6-9d86e0aaa810-serviceca\") pod \"node-ca-gdgdx\" (UID: \"5d82368b-ed78-4142-b1d6-9d86e0aaa810\") " pod="openshift-image-registry/node-ca-gdgdx" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.915799 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.918248 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.918282 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.918291 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.918306 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.918315 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:40Z","lastTransitionTime":"2026-03-17T15:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.930049 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqd4h\" (UniqueName: \"kubernetes.io/projected/5d82368b-ed78-4142-b1d6-9d86e0aaa810-kube-api-access-cqd4h\") pod \"node-ca-gdgdx\" (UID: \"5d82368b-ed78-4142-b1d6-9d86e0aaa810\") " pod="openshift-image-registry/node-ca-gdgdx" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.932146 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:40 crc kubenswrapper[4767]: I0317 15:38:40.943329 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:40Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.021742 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.021791 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.021806 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.021830 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.021848 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:41Z","lastTransitionTime":"2026-03-17T15:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.041910 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gdgdx" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.111122 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gdgdx" event={"ID":"5d82368b-ed78-4142-b1d6-9d86e0aaa810","Type":"ContainerStarted","Data":"ee5cda68413b6c7eba32167f2a2fc650efdf836c90011211065538b56f1f90af"} Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.116034 4767 generic.go:334] "Generic (PLEG): container finished" podID="811a8833-5459-4af6-a1c3-179040eb3c01" containerID="68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a" exitCode=0 Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.116086 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" event={"ID":"811a8833-5459-4af6-a1c3-179040eb3c01","Type":"ContainerDied","Data":"68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a"} Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.132475 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.132526 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.132541 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.132564 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.132580 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:41Z","lastTransitionTime":"2026-03-17T15:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.132983 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.148145 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.162196 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.176551 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.188849 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.199021 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.216399 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.228123 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.234666 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.234702 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.234714 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.234727 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.234738 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:41Z","lastTransitionTime":"2026-03-17T15:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.239595 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.258083 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.272342 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.283850 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.298501 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.310073 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:41Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.337284 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.337328 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.337341 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.337376 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.337395 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:41Z","lastTransitionTime":"2026-03-17T15:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.439918 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.439959 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.439968 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.439983 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.439992 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:41Z","lastTransitionTime":"2026-03-17T15:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.542246 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.542299 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.542312 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.542334 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.542349 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:41Z","lastTransitionTime":"2026-03-17T15:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.645154 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.645230 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.645250 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.645281 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.645300 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:41Z","lastTransitionTime":"2026-03-17T15:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.747422 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.747467 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.747484 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.747514 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.747532 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:41Z","lastTransitionTime":"2026-03-17T15:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.853761 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.853854 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.853872 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.853896 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.853924 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:41Z","lastTransitionTime":"2026-03-17T15:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.956533 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.956569 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.956580 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.956593 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:41 crc kubenswrapper[4767]: I0317 15:38:41.956602 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:41Z","lastTransitionTime":"2026-03-17T15:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.058842 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.058878 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.058886 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.058898 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.058907 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:42Z","lastTransitionTime":"2026-03-17T15:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.126158 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.127877 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.127954 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.128045 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.137270 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" event={"ID":"811a8833-5459-4af6-a1c3-179040eb3c01","Type":"ContainerStarted","Data":"747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.139962 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gdgdx" event={"ID":"5d82368b-ed78-4142-b1d6-9d86e0aaa810","Type":"ContainerStarted","Data":"6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.141780 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.223158 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.223421 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.223439 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.223446 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.223459 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.223467 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:42Z","lastTransitionTime":"2026-03-17T15:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.258412 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.258495 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.267276 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.278942 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.294578 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.311536 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.325544 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.325871 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.325885 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.325900 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.325911 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:42Z","lastTransitionTime":"2026-03-17T15:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.331244 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.354552 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.354639 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:42 crc kubenswrapper[4767]: E0317 15:38:42.354690 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.354552 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:42 crc kubenswrapper[4767]: E0317 15:38:42.354775 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:42 crc kubenswrapper[4767]: E0317 15:38:42.354823 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.356448 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.370783 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.383053 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.398742 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.414361 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.430674 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.430729 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.430744 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.430768 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.430784 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:42Z","lastTransitionTime":"2026-03-17T15:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.432585 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.453865 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.475194 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.491533 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.507923 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.525087 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.533825 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.534104 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.534223 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.534323 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.534401 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:42Z","lastTransitionTime":"2026-03-17T15:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.542527 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.567743 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.586883 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.605008 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.622160 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.634356 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.636330 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.636364 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.636377 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.636393 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.636438 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:42Z","lastTransitionTime":"2026-03-17T15:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.647577 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.661137 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.675615 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.696194 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:42Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.739329 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.739652 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.739761 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.739871 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.739956 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:42Z","lastTransitionTime":"2026-03-17T15:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.842375 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.842416 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.842428 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.842443 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.842453 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:42Z","lastTransitionTime":"2026-03-17T15:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.946151 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.946241 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.946267 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.946295 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:42 crc kubenswrapper[4767]: I0317 15:38:42.946313 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:42Z","lastTransitionTime":"2026-03-17T15:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.072870 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.072910 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.072925 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.072950 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.072964 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.260411 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.260478 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.260518 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.260541 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.260558 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.363069 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.363123 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.363136 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.363152 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.363163 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.365999 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.465722 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.465772 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.465785 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.465803 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.465817 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.568856 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.568906 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.568922 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.568944 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.568956 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.671477 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.671576 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.671593 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.671955 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.672032 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.774209 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.774236 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.774244 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.774256 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.774265 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.877916 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.877953 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.877962 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.877974 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.877982 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.936025 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.936089 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.936107 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.936129 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.936146 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: E0317 15:38:43.954900 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:43Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.959590 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.959636 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.959651 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.959666 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.959678 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: E0317 15:38:43.973799 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:43Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.978964 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.979017 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.979036 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.979060 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.979077 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:43 crc kubenswrapper[4767]: E0317 15:38:43.992044 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:43Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.996853 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.996978 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.997075 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.997192 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:43 crc kubenswrapper[4767]: I0317 15:38:43.997291 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:43Z","lastTransitionTime":"2026-03-17T15:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:44 crc kubenswrapper[4767]: E0317 15:38:44.011523 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.017147 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.017230 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.017251 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.017271 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.017287 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:44Z","lastTransitionTime":"2026-03-17T15:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:44 crc kubenswrapper[4767]: E0317 15:38:44.033616 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: E0317 15:38:44.033901 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.036016 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.036087 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.036114 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.036148 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.036239 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:44Z","lastTransitionTime":"2026-03-17T15:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.139050 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.139098 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.139113 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.139145 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.139163 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:44Z","lastTransitionTime":"2026-03-17T15:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.242267 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.242303 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.242313 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.242351 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.242364 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:44Z","lastTransitionTime":"2026-03-17T15:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.268495 4767 generic.go:334] "Generic (PLEG): container finished" podID="811a8833-5459-4af6-a1c3-179040eb3c01" containerID="747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c" exitCode=0 Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.268564 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" event={"ID":"811a8833-5459-4af6-a1c3-179040eb3c01","Type":"ContainerDied","Data":"747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c"} Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.301837 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.320158 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.335772 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.347807 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.348030 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.348041 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.348063 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.348073 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:44Z","lastTransitionTime":"2026-03-17T15:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.350585 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.354278 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.354317 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:44 crc kubenswrapper[4767]: E0317 15:38:44.354396 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.354696 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:44 crc kubenswrapper[4767]: E0317 15:38:44.354758 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:44 crc kubenswrapper[4767]: E0317 15:38:44.354804 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.366588 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.377493 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.438475 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.450302 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.450934 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.450980 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.450991 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.451005 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.451015 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:44Z","lastTransitionTime":"2026-03-17T15:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.459964 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.473018 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.484180 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.499022 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.510707 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.521358 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.539001 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:44Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.554041 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.554092 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.554102 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.554122 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.554134 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:44Z","lastTransitionTime":"2026-03-17T15:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.657000 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.657060 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.657083 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.657115 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.657136 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:44Z","lastTransitionTime":"2026-03-17T15:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.759451 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.759505 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.759519 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.759539 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.759554 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:44Z","lastTransitionTime":"2026-03-17T15:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.862307 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.862366 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.862386 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.862432 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.862457 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:44Z","lastTransitionTime":"2026-03-17T15:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.965213 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.965275 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.965292 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.965314 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:44 crc kubenswrapper[4767]: I0317 15:38:44.965333 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:44Z","lastTransitionTime":"2026-03-17T15:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.068606 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.068648 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.068659 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.068688 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.068699 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:45Z","lastTransitionTime":"2026-03-17T15:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.171123 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.171251 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.171283 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.171311 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.171329 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:45Z","lastTransitionTime":"2026-03-17T15:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.273806 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.273840 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.273848 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.273862 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.273870 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:45Z","lastTransitionTime":"2026-03-17T15:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.285616 4767 generic.go:334] "Generic (PLEG): container finished" podID="811a8833-5459-4af6-a1c3-179040eb3c01" containerID="d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06" exitCode=0 Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.285650 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" event={"ID":"811a8833-5459-4af6-a1c3-179040eb3c01","Type":"ContainerDied","Data":"d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06"} Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.308842 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.324832 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.376080 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.376138 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.376150 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.376170 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.376207 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:45Z","lastTransitionTime":"2026-03-17T15:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.479298 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.479356 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.479371 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.479388 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.479402 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:45Z","lastTransitionTime":"2026-03-17T15:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.509701 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.520852 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.537487 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.560195 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.573772 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.581842 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.581872 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.581897 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.581910 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.581918 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:45Z","lastTransitionTime":"2026-03-17T15:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.587456 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.599303 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.612464 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.622715 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.633496 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.648126 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.662489 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.676856 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:45Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.685094 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.685518 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.685544 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.685568 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.685846 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:45Z","lastTransitionTime":"2026-03-17T15:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.789601 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.789703 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.789720 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.789758 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.789776 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:45Z","lastTransitionTime":"2026-03-17T15:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.893107 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.893615 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.893866 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.894090 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.894480 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:45Z","lastTransitionTime":"2026-03-17T15:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.997970 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.998031 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.998050 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.998072 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:45 crc kubenswrapper[4767]: I0317 15:38:45.998086 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:45Z","lastTransitionTime":"2026-03-17T15:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.100366 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.100668 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.100748 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.100852 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.100931 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:46Z","lastTransitionTime":"2026-03-17T15:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.268351 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.268385 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.268395 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.268411 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.268420 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:46Z","lastTransitionTime":"2026-03-17T15:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.293666 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" event={"ID":"811a8833-5459-4af6-a1c3-179040eb3c01","Type":"ContainerStarted","Data":"1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175"} Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.329320 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.454422 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:46 crc kubenswrapper[4767]: E0317 15:38:46.454819 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.454520 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:46 crc kubenswrapper[4767]: E0317 15:38:46.455083 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.454492 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:46 crc kubenswrapper[4767]: E0317 15:38:46.455320 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.456700 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.456727 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.456735 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.456745 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.456753 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:46Z","lastTransitionTime":"2026-03-17T15:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.468495 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.486875 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.519581 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.534140 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.545011 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.558304 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.558742 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.558792 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.558804 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.558820 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.558830 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:46Z","lastTransitionTime":"2026-03-17T15:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.571062 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.580208 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.591043 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.603987 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.617907 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.632470 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.641848 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.660952 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:46Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.661740 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.661783 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.661792 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.661807 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.661816 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:46Z","lastTransitionTime":"2026-03-17T15:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.879829 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.880044 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.880055 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.880071 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.880084 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:46Z","lastTransitionTime":"2026-03-17T15:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.981560 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.981597 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.981609 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.981624 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:46 crc kubenswrapper[4767]: I0317 15:38:46.981635 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:46Z","lastTransitionTime":"2026-03-17T15:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.084645 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.084680 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.084688 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.084701 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.084710 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:47Z","lastTransitionTime":"2026-03-17T15:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.187353 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.187403 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.187414 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.187431 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.187440 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:47Z","lastTransitionTime":"2026-03-17T15:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.270783 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr"] Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.271466 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.273816 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.273924 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.289060 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.289130 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.289143 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.289163 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.289188 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:47Z","lastTransitionTime":"2026-03-17T15:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.290068 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.305316 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.317446 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.328196 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.339744 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.352186 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.378164 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.387768 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c277ffee-5af0-4b9e-8afc-206876e483b0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.387807 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c277ffee-5af0-4b9e-8afc-206876e483b0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.387828 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c277ffee-5af0-4b9e-8afc-206876e483b0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.387875 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-845s5\" (UniqueName: \"kubernetes.io/projected/c277ffee-5af0-4b9e-8afc-206876e483b0-kube-api-access-845s5\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.391295 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.391348 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.391362 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.391381 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.391395 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:47Z","lastTransitionTime":"2026-03-17T15:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.394148 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.414467 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.432778 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.448104 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.461455 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.473358 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.485395 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.488577 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-845s5\" (UniqueName: \"kubernetes.io/projected/c277ffee-5af0-4b9e-8afc-206876e483b0-kube-api-access-845s5\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.488648 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c277ffee-5af0-4b9e-8afc-206876e483b0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.488679 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c277ffee-5af0-4b9e-8afc-206876e483b0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.488712 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c277ffee-5af0-4b9e-8afc-206876e483b0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.489841 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c277ffee-5af0-4b9e-8afc-206876e483b0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.490475 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c277ffee-5af0-4b9e-8afc-206876e483b0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.493566 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.493603 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.493615 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.493631 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.493643 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:47Z","lastTransitionTime":"2026-03-17T15:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.496071 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c277ffee-5af0-4b9e-8afc-206876e483b0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.501089 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.506851 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-845s5\" (UniqueName: \"kubernetes.io/projected/c277ffee-5af0-4b9e-8afc-206876e483b0-kube-api-access-845s5\") pod \"ovnkube-control-plane-749d76644c-bprkr\" (UID: \"c277ffee-5af0-4b9e-8afc-206876e483b0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.511529 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:47Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.590971 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.595928 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.595980 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.595993 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.596010 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.596021 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:47Z","lastTransitionTime":"2026-03-17T15:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:47 crc kubenswrapper[4767]: W0317 15:38:47.605856 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc277ffee_5af0_4b9e_8afc_206876e483b0.slice/crio-eb80347bc67a37f423904938c31a8a2650dbd55eaea5bbac74b19adece99467d WatchSource:0}: Error finding container eb80347bc67a37f423904938c31a8a2650dbd55eaea5bbac74b19adece99467d: Status 404 returned error can't find the container with id eb80347bc67a37f423904938c31a8a2650dbd55eaea5bbac74b19adece99467d Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.779364 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.779427 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.779439 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.779468 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.779479 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:47Z","lastTransitionTime":"2026-03-17T15:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.881402 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.881437 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.881445 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.881457 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.881465 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:47Z","lastTransitionTime":"2026-03-17T15:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.983531 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.983584 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.983595 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.983607 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:47 crc kubenswrapper[4767]: I0317 15:38:47.983615 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:47Z","lastTransitionTime":"2026-03-17T15:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.061799 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-t7mfn"] Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.062267 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:48 crc kubenswrapper[4767]: E0317 15:38:48.062316 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.085485 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.085545 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.085565 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.085590 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.085606 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:48Z","lastTransitionTime":"2026-03-17T15:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.095707 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.113489 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.131033 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.143833 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.157159 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.171147 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.183354 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.183437 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6xc7\" (UniqueName: \"kubernetes.io/projected/32925427-1529-4e8d-a606-8ccbde1fd877-kube-api-access-b6xc7\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.185672 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.187381 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.187414 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.187424 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.187447 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.187456 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:48Z","lastTransitionTime":"2026-03-17T15:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.203190 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.212371 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.222361 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.233840 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.243503 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.254894 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.265410 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.279767 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.284728 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6xc7\" (UniqueName: \"kubernetes.io/projected/32925427-1529-4e8d-a606-8ccbde1fd877-kube-api-access-b6xc7\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.284805 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:48 crc kubenswrapper[4767]: E0317 15:38:48.284981 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:38:48 crc kubenswrapper[4767]: E0317 15:38:48.285047 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs podName:32925427-1529-4e8d-a606-8ccbde1fd877 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:48.785031264 +0000 UTC m=+120.198347311 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs") pod "network-metrics-daemon-t7mfn" (UID: "32925427-1529-4e8d-a606-8ccbde1fd877") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.289657 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.289706 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.289715 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.289727 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.289738 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:48Z","lastTransitionTime":"2026-03-17T15:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.292332 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.300282 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovnkube-controller/0.log" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.302812 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6xc7\" (UniqueName: \"kubernetes.io/projected/32925427-1529-4e8d-a606-8ccbde1fd877-kube-api-access-b6xc7\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.303472 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a" exitCode=1 Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.303554 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.304271 4767 scope.go:117] "RemoveContainer" containerID="a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.305406 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" event={"ID":"c277ffee-5af0-4b9e-8afc-206876e483b0","Type":"ContainerStarted","Data":"b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.305475 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" event={"ID":"c277ffee-5af0-4b9e-8afc-206876e483b0","Type":"ContainerStarted","Data":"ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.305498 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" event={"ID":"c277ffee-5af0-4b9e-8afc-206876e483b0","Type":"ContainerStarted","Data":"eb80347bc67a37f423904938c31a8a2650dbd55eaea5bbac74b19adece99467d"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.310537 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.322422 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.335666 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.347725 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.353426 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.353477 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.353526 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:48 crc kubenswrapper[4767]: E0317 15:38:48.353539 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:48 crc kubenswrapper[4767]: E0317 15:38:48.353619 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:48 crc kubenswrapper[4767]: E0317 15:38:48.353763 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.367418 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.376653 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.392735 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.395039 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.395076 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.395085 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.395102 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.395116 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:48Z","lastTransitionTime":"2026-03-17T15:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.407347 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.419843 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.431927 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.444717 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.457477 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.484713 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"message\\\":\\\"4\\\\nI0317 15:38:47.835219 6527 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:47.835259 6527 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:47.835266 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:47.835326 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:47.835340 6527 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:47.835362 6527 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:47.835358 6527 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:47.835383 6527 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:38:47.835398 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:47.835408 6527 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:47.835413 6527 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:47.835418 6527 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:38:47.835434 6527 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:38:47.835443 6527 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:47.835494 6527 factory.go:656] Stopping watch factory\\\\nI0317 15:38:47.835510 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.506386 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.508418 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.508452 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.508464 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.508479 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.508488 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:48Z","lastTransitionTime":"2026-03-17T15:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.529458 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.544028 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.557711 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.569280 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:48Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.614307 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.614362 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.614375 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.614391 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.614403 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:48Z","lastTransitionTime":"2026-03-17T15:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.717121 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.717151 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.717159 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.717194 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.717203 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:48Z","lastTransitionTime":"2026-03-17T15:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.788165 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:48 crc kubenswrapper[4767]: E0317 15:38:48.788356 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:38:48 crc kubenswrapper[4767]: E0317 15:38:48.788465 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs podName:32925427-1529-4e8d-a606-8ccbde1fd877 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:49.788439447 +0000 UTC m=+121.201755514 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs") pod "network-metrics-daemon-t7mfn" (UID: "32925427-1529-4e8d-a606-8ccbde1fd877") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.820357 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.820417 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.820432 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.820481 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.820506 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:48Z","lastTransitionTime":"2026-03-17T15:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.922989 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.923065 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.923105 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.923140 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:48 crc kubenswrapper[4767]: I0317 15:38:48.923155 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:48Z","lastTransitionTime":"2026-03-17T15:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.026464 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.026498 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.026507 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.026520 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.026530 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:49Z","lastTransitionTime":"2026-03-17T15:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.153137 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.153202 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.153215 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.153231 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.153243 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:49Z","lastTransitionTime":"2026-03-17T15:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.255868 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.255910 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.255920 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.255935 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.255943 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:49Z","lastTransitionTime":"2026-03-17T15:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.311231 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovnkube-controller/0.log" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.313410 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572"} Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.313771 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.325816 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.341869 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.353290 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:49 crc kubenswrapper[4767]: E0317 15:38:49.353429 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:38:49 crc kubenswrapper[4767]: E0317 15:38:49.356103 4767 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.361157 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.375058 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.388630 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.400620 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.423742 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"message\\\":\\\"4\\\\nI0317 15:38:47.835219 6527 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:47.835259 6527 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:47.835266 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:47.835326 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:47.835340 6527 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:47.835362 6527 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:47.835358 6527 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:47.835383 6527 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:38:47.835398 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:47.835408 6527 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:47.835413 6527 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:47.835418 6527 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:38:47.835434 6527 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:38:47.835443 6527 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:47.835494 6527 factory.go:656] Stopping watch factory\\\\nI0317 15:38:47.835510 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.437438 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.460421 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.480082 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.501596 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.516617 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.531959 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.549494 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.574809 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.591185 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.602682 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.618718 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: E0317 15:38:49.621038 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.631462 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.642834 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.652132 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.669730 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"message\\\":\\\"4\\\\nI0317 15:38:47.835219 6527 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:47.835259 6527 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:47.835266 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:47.835326 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:47.835340 6527 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:47.835362 6527 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:47.835358 6527 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:47.835383 6527 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:38:47.835398 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:47.835408 6527 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:47.835413 6527 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:47.835418 6527 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:38:47.835434 6527 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:38:47.835443 6527 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:47.835494 6527 factory.go:656] Stopping watch factory\\\\nI0317 15:38:47.835510 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.682552 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.701285 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.713785 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.726628 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.738802 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.751624 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.765593 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.778484 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.792338 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.796580 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:49 crc kubenswrapper[4767]: E0317 15:38:49.796729 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:38:49 crc kubenswrapper[4767]: E0317 15:38:49.796815 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs podName:32925427-1529-4e8d-a606-8ccbde1fd877 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:51.796780447 +0000 UTC m=+123.210096504 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs") pod "network-metrics-daemon-t7mfn" (UID: "32925427-1529-4e8d-a606-8ccbde1fd877") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.802824 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.813818 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:49 crc kubenswrapper[4767]: I0317 15:38:49.825855 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:49Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.319966 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovnkube-controller/1.log" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.321017 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovnkube-controller/0.log" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.324992 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572" exitCode=1 Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.325074 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572"} Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.325132 4767 scope.go:117] "RemoveContainer" containerID="a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.326448 4767 scope.go:117] "RemoveContainer" containerID="92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572" Mar 17 15:38:50 crc kubenswrapper[4767]: E0317 15:38:50.327083 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.343588 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.353367 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.353442 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.353375 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:50 crc kubenswrapper[4767]: E0317 15:38:50.353620 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:50 crc kubenswrapper[4767]: E0317 15:38:50.353776 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:50 crc kubenswrapper[4767]: E0317 15:38:50.353887 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.428772 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.442463 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.453146 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.461057 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.469846 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.487747 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.499537 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.513504 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.530743 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.541965 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.558212 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.571670 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.584799 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.612151 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"message\\\":\\\"4\\\\nI0317 15:38:47.835219 6527 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:47.835259 6527 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:47.835266 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:47.835326 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:47.835340 6527 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:47.835362 6527 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:47.835358 6527 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:47.835383 6527 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:38:47.835398 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:47.835408 6527 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:47.835413 6527 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:47.835418 6527 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:38:47.835434 6527 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:38:47.835443 6527 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:47.835494 6527 factory.go:656] Stopping watch factory\\\\nI0317 15:38:47.835510 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:49Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:49.898398 6812 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:38:49.898411 6812 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:49.898421 6812 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:49.898431 6812 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 15:38:49.898446 6812 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:49.898459 6812 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:49.898439 6812 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:49.898569 6812 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:49.898579 6812 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:49.898623 6812 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:49.898662 6812 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:38:49.898705 6812 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:49.899029 6812 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:49.899082 6812 factory.go:656] Stopping watch factory\\\\nI0317 15:38:49.899100 6812 ovnkube.go:599] Stopped ovnkube\\\\nI0317 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.625408 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.644584 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.658212 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.668071 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.689581 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.702874 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.717299 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.739143 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.753898 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.767770 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.779587 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.793826 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.805028 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.819509 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.830877 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.840114 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.856993 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0312aeb6206439984ae92fe147f4dbf80062cc5cc9884e81b48804590a0156a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"message\\\":\\\"4\\\\nI0317 15:38:47.835219 6527 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:47.835259 6527 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:47.835266 6527 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:47.835326 6527 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:47.835340 6527 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:47.835362 6527 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:47.835358 6527 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:47.835383 6527 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:38:47.835398 6527 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:47.835408 6527 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:47.835413 6527 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:47.835418 6527 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:38:47.835434 6527 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:38:47.835443 6527 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:47.835494 6527 factory.go:656] Stopping watch factory\\\\nI0317 15:38:47.835510 6527 ovnkube.go:599] Stopped ovnkube\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:49Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:49.898398 6812 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:38:49.898411 6812 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:49.898421 6812 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:49.898431 6812 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 15:38:49.898446 6812 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:49.898459 6812 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:49.898439 6812 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:49.898569 6812 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:49.898579 6812 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:49.898623 6812 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:49.898662 6812 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:38:49.898705 6812 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:49.899029 6812 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:49.899082 6812 factory.go:656] Stopping watch factory\\\\nI0317 15:38:49.899100 6812 ovnkube.go:599] Stopped ovnkube\\\\nI0317 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.867481 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.881537 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:50 crc kubenswrapper[4767]: I0317 15:38:50.892923 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:50Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.329390 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovnkube-controller/1.log" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.336589 4767 scope.go:117] "RemoveContainer" containerID="92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572" Mar 17 15:38:51 crc kubenswrapper[4767]: E0317 15:38:51.337453 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.352888 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.353304 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:51 crc kubenswrapper[4767]: E0317 15:38:51.353462 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.487130 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.499405 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.514924 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.528394 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.539693 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.557303 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:49Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:49.898398 6812 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:38:49.898411 6812 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:49.898421 6812 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:49.898431 6812 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 15:38:49.898446 6812 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:49.898459 6812 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:49.898439 6812 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:49.898569 6812 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:49.898579 6812 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:49.898623 6812 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:49.898662 6812 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:38:49.898705 6812 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:49.899029 6812 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:49.899082 6812 factory.go:656] Stopping watch factory\\\\nI0317 15:38:49.899100 6812 ovnkube.go:599] Stopped ovnkube\\\\nI0317 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.569994 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.583780 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.595507 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.605646 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.625118 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.638565 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.651106 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.667984 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.677210 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.686430 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:51Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.888188 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:51 crc kubenswrapper[4767]: E0317 15:38:51.888360 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:38:51 crc kubenswrapper[4767]: E0317 15:38:51.888410 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs podName:32925427-1529-4e8d-a606-8ccbde1fd877 nodeName:}" failed. No retries permitted until 2026-03-17 15:38:55.888395879 +0000 UTC m=+127.301711926 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs") pod "network-metrics-daemon-t7mfn" (UID: "32925427-1529-4e8d-a606-8ccbde1fd877") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:38:51 crc kubenswrapper[4767]: I0317 15:38:51.913128 4767 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 17 15:38:52 crc kubenswrapper[4767]: I0317 15:38:52.354048 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:52 crc kubenswrapper[4767]: I0317 15:38:52.354154 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:52 crc kubenswrapper[4767]: I0317 15:38:52.354066 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:52 crc kubenswrapper[4767]: E0317 15:38:52.354401 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:52 crc kubenswrapper[4767]: E0317 15:38:52.354497 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:52 crc kubenswrapper[4767]: E0317 15:38:52.354600 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:53 crc kubenswrapper[4767]: I0317 15:38:53.353630 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:53 crc kubenswrapper[4767]: E0317 15:38:53.353769 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.244448 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.244502 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.244519 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.244542 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.244559 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:54Z","lastTransitionTime":"2026-03-17T15:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:54 crc kubenswrapper[4767]: E0317 15:38:54.263818 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:54Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.269506 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.269578 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.269603 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.269632 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.269656 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:54Z","lastTransitionTime":"2026-03-17T15:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:54 crc kubenswrapper[4767]: E0317 15:38:54.292502 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:54Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.297574 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.297644 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.297663 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.297688 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.297708 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:54Z","lastTransitionTime":"2026-03-17T15:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:54 crc kubenswrapper[4767]: E0317 15:38:54.321046 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:54Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.326656 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.326728 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.326750 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.326776 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.326795 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:54Z","lastTransitionTime":"2026-03-17T15:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:54 crc kubenswrapper[4767]: E0317 15:38:54.344383 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:54Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.349383 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.349436 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.349448 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.349466 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.349482 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:38:54Z","lastTransitionTime":"2026-03-17T15:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.353871 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.353900 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:54 crc kubenswrapper[4767]: I0317 15:38:54.353891 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:54 crc kubenswrapper[4767]: E0317 15:38:54.354035 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:54 crc kubenswrapper[4767]: E0317 15:38:54.354132 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:54 crc kubenswrapper[4767]: E0317 15:38:54.354372 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:54 crc kubenswrapper[4767]: E0317 15:38:54.369241 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:54Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:54 crc kubenswrapper[4767]: E0317 15:38:54.369566 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 15:38:54 crc kubenswrapper[4767]: E0317 15:38:54.621768 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:38:55 crc kubenswrapper[4767]: I0317 15:38:55.353449 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:55 crc kubenswrapper[4767]: E0317 15:38:55.353680 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:38:55 crc kubenswrapper[4767]: I0317 15:38:55.928117 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:55 crc kubenswrapper[4767]: E0317 15:38:55.928323 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:38:55 crc kubenswrapper[4767]: E0317 15:38:55.928418 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs podName:32925427-1529-4e8d-a606-8ccbde1fd877 nodeName:}" failed. No retries permitted until 2026-03-17 15:39:03.928392436 +0000 UTC m=+135.341708513 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs") pod "network-metrics-daemon-t7mfn" (UID: "32925427-1529-4e8d-a606-8ccbde1fd877") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:38:56 crc kubenswrapper[4767]: I0317 15:38:56.130451 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.130587 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:28.130567339 +0000 UTC m=+159.543883396 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:38:56 crc kubenswrapper[4767]: I0317 15:38:56.130746 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:56 crc kubenswrapper[4767]: I0317 15:38:56.130812 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.130951 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.131046 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:39:28.131026802 +0000 UTC m=+159.544342849 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.131046 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.131295 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:39:28.131281429 +0000 UTC m=+159.544597486 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:38:56 crc kubenswrapper[4767]: I0317 15:38:56.231766 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:56 crc kubenswrapper[4767]: I0317 15:38:56.231848 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.231936 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.231972 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.231987 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.232004 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.232027 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.232047 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.232055 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 15:39:28.232036846 +0000 UTC m=+159.645352903 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.232151 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 15:39:28.232131419 +0000 UTC m=+159.645447496 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:38:56 crc kubenswrapper[4767]: I0317 15:38:56.353428 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:56 crc kubenswrapper[4767]: I0317 15:38:56.353473 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:56 crc kubenswrapper[4767]: I0317 15:38:56.353506 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.353572 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.353720 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:56 crc kubenswrapper[4767]: E0317 15:38:56.353806 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:57 crc kubenswrapper[4767]: I0317 15:38:57.354281 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:57 crc kubenswrapper[4767]: E0317 15:38:57.354505 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:38:58 crc kubenswrapper[4767]: I0317 15:38:58.353378 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:38:58 crc kubenswrapper[4767]: I0317 15:38:58.353458 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:38:58 crc kubenswrapper[4767]: E0317 15:38:58.353501 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:38:58 crc kubenswrapper[4767]: I0317 15:38:58.353576 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:38:58 crc kubenswrapper[4767]: E0317 15:38:58.353656 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:38:58 crc kubenswrapper[4767]: E0317 15:38:58.353758 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.354112 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:38:59 crc kubenswrapper[4767]: E0317 15:38:59.354367 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.368924 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.381716 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.396378 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.413738 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.427877 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.443005 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.500977 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.531585 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:49Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:49.898398 6812 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:38:49.898411 6812 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:49.898421 6812 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:49.898431 6812 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 15:38:49.898446 6812 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:49.898459 6812 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:49.898439 6812 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:49.898569 6812 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:49.898579 6812 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:49.898623 6812 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:49.898662 6812 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:38:49.898705 6812 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:49.899029 6812 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:49.899082 6812 factory.go:656] Stopping watch factory\\\\nI0317 15:38:49.899100 6812 ovnkube.go:599] Stopped ovnkube\\\\nI0317 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.549829 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.567477 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.584995 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.595517 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.609759 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: E0317 15:38:59.623052 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.626770 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.643559 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.662477 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:38:59 crc kubenswrapper[4767]: I0317 15:38:59.675297 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:38:59Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:00 crc kubenswrapper[4767]: I0317 15:39:00.353766 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:00 crc kubenswrapper[4767]: I0317 15:39:00.353811 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:00 crc kubenswrapper[4767]: I0317 15:39:00.353825 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:00 crc kubenswrapper[4767]: E0317 15:39:00.353933 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:00 crc kubenswrapper[4767]: E0317 15:39:00.354101 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:00 crc kubenswrapper[4767]: E0317 15:39:00.354259 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:01 crc kubenswrapper[4767]: I0317 15:39:01.353494 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:01 crc kubenswrapper[4767]: E0317 15:39:01.354071 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:02 crc kubenswrapper[4767]: I0317 15:39:02.353807 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:02 crc kubenswrapper[4767]: E0317 15:39:02.353946 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:02 crc kubenswrapper[4767]: I0317 15:39:02.353807 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:02 crc kubenswrapper[4767]: I0317 15:39:02.353811 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:02 crc kubenswrapper[4767]: E0317 15:39:02.354429 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:02 crc kubenswrapper[4767]: E0317 15:39:02.354243 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:03 crc kubenswrapper[4767]: I0317 15:39:03.353894 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:03 crc kubenswrapper[4767]: E0317 15:39:03.354153 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.011973 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.012101 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.012148 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs podName:32925427-1529-4e8d-a606-8ccbde1fd877 nodeName:}" failed. No retries permitted until 2026-03-17 15:39:20.012136039 +0000 UTC m=+151.425452086 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs") pod "network-metrics-daemon-t7mfn" (UID: "32925427-1529-4e8d-a606-8ccbde1fd877") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.354308 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.354379 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.354427 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.354558 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.354716 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.354804 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.624821 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.686128 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.686224 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.686239 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.686256 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.686270 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:04Z","lastTransitionTime":"2026-03-17T15:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.707617 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:04Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.712676 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.712779 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.712799 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.712823 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.712877 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:04Z","lastTransitionTime":"2026-03-17T15:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.734026 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:04Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.739278 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.739331 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.739348 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.739373 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.739389 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:04Z","lastTransitionTime":"2026-03-17T15:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.756338 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:04Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.761842 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.761907 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.761949 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.761985 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.762010 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:04Z","lastTransitionTime":"2026-03-17T15:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.836550 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:04Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.840268 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.840304 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.840314 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.840328 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:04 crc kubenswrapper[4767]: I0317 15:39:04.840337 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:04Z","lastTransitionTime":"2026-03-17T15:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.851662 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:04Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:04 crc kubenswrapper[4767]: E0317 15:39:04.851788 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 15:39:05 crc kubenswrapper[4767]: I0317 15:39:05.353509 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:05 crc kubenswrapper[4767]: E0317 15:39:05.353741 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:06 crc kubenswrapper[4767]: I0317 15:39:06.353430 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:06 crc kubenswrapper[4767]: I0317 15:39:06.353503 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:06 crc kubenswrapper[4767]: E0317 15:39:06.353860 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:06 crc kubenswrapper[4767]: I0317 15:39:06.353430 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:06 crc kubenswrapper[4767]: E0317 15:39:06.354085 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:06 crc kubenswrapper[4767]: E0317 15:39:06.354876 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:06 crc kubenswrapper[4767]: I0317 15:39:06.355441 4767 scope.go:117] "RemoveContainer" containerID="92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572" Mar 17 15:39:06 crc kubenswrapper[4767]: I0317 15:39:06.378592 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.470374 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:07 crc kubenswrapper[4767]: E0317 15:39:07.470670 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.483636 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovnkube-controller/1.log" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.486270 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b"} Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.487232 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.627341 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.641203 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.653272 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.664208 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.681525 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.691115 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f49ff1ec-7571-476d-be04-40c50ff291d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2979605c7105c9f289cbfc223356236cea8fb72fec0b0242dea17fb1ad10e950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.704936 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.721574 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.734703 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.744924 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.754676 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.768756 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.780369 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.795769 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.810029 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.821698 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.831692 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:07 crc kubenswrapper[4767]: I0317 15:39:07.849531 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:49Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:49.898398 6812 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:38:49.898411 6812 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:49.898421 6812 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:49.898431 6812 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 15:38:49.898446 6812 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:49.898459 6812 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:49.898439 6812 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:49.898569 6812 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:49.898579 6812 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:49.898623 6812 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:49.898662 6812 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:38:49.898705 6812 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:49.899029 6812 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:49.899082 6812 factory.go:656] Stopping watch factory\\\\nI0317 15:38:49.899100 6812 ovnkube.go:599] Stopped ovnkube\\\\nI0317 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:07Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.353629 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:08 crc kubenswrapper[4767]: E0317 15:39:08.353964 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.353695 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:08 crc kubenswrapper[4767]: E0317 15:39:08.354035 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.353668 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:08 crc kubenswrapper[4767]: E0317 15:39:08.354083 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.491822 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovnkube-controller/2.log" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.493249 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovnkube-controller/1.log" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.495924 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b" exitCode=1 Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.495963 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b"} Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.496009 4767 scope.go:117] "RemoveContainer" containerID="92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.496657 4767 scope.go:117] "RemoveContainer" containerID="544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b" Mar 17 15:39:08 crc kubenswrapper[4767]: E0317 15:39:08.496797 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.514305 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.528855 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.557279 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.573697 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.591342 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.602988 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.624825 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.638837 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.658234 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.670122 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f49ff1ec-7571-476d-be04-40c50ff291d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2979605c7105c9f289cbfc223356236cea8fb72fec0b0242dea17fb1ad10e950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.682222 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.698029 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.713937 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.730388 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.743951 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.765504 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:49Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:49.898398 6812 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:38:49.898411 6812 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:49.898421 6812 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:49.898431 6812 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 15:38:49.898446 6812 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:49.898459 6812 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:49.898439 6812 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:49.898569 6812 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:49.898579 6812 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:49.898623 6812 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:49.898662 6812 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:38:49.898705 6812 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:49.899029 6812 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:49.899082 6812 factory.go:656] Stopping watch factory\\\\nI0317 15:38:49.899100 6812 ovnkube.go:599] Stopped ovnkube\\\\nI0317 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:39:07Z\\\",\\\"message\\\":\\\" 15:39:07.862953 7018 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:39:07.863029 7018 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:39:07.863040 7018 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:39:07.863068 7018 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:39:07.863285 7018 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:39:07.863298 7018 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:39:07.863304 7018 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:39:07.863317 7018 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:39:07.863407 7018 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:39:07.863536 7018 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 15:39:07.863581 7018 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:39:07.863589 7018 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:39:07.863604 7018 factory.go:656] Stopping watch factory\\\\nI0317 15:39:07.863620 7018 ovnkube.go:599] Stopped ovnkube\\\\nI0317 15:39:07.863678 7018 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:39:07.863676 7018 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0317 15\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.780364 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:08 crc kubenswrapper[4767]: I0317 15:39:08.794706 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:08Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.354109 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:09 crc kubenswrapper[4767]: E0317 15:39:09.354239 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.370421 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.385023 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.399872 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.412556 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.430440 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ced2d889074c8bb8e50b7bfe77479eaa8c52573f8d0c80db8ba7d40140d572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:38:49Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:38:49.898398 6812 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:38:49.898411 6812 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 15:38:49.898421 6812 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 15:38:49.898431 6812 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 15:38:49.898446 6812 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 15:38:49.898459 6812 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:38:49.898439 6812 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:38:49.898569 6812 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:38:49.898579 6812 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:38:49.898623 6812 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:38:49.898662 6812 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:38:49.898705 6812 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:38:49.899029 6812 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 15:38:49.899082 6812 factory.go:656] Stopping watch factory\\\\nI0317 15:38:49.899100 6812 ovnkube.go:599] Stopped ovnkube\\\\nI0317 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:39:07Z\\\",\\\"message\\\":\\\" 15:39:07.862953 7018 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:39:07.863029 7018 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:39:07.863040 7018 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:39:07.863068 7018 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:39:07.863285 7018 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:39:07.863298 7018 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:39:07.863304 7018 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:39:07.863317 7018 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:39:07.863407 7018 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:39:07.863536 7018 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 15:39:07.863581 7018 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:39:07.863589 7018 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:39:07.863604 7018 factory.go:656] Stopping watch factory\\\\nI0317 15:39:07.863620 7018 ovnkube.go:599] Stopped ovnkube\\\\nI0317 15:39:07.863678 7018 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:39:07.863676 7018 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0317 15\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.441548 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.462741 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.475259 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.492816 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.500419 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovnkube-controller/2.log" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.504331 4767 scope.go:117] "RemoveContainer" containerID="544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b" Mar 17 15:39:09 crc kubenswrapper[4767]: E0317 15:39:09.504486 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.504781 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.515887 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.525313 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f49ff1ec-7571-476d-be04-40c50ff291d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2979605c7105c9f289cbfc223356236cea8fb72fec0b0242dea17fb1ad10e950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.539438 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.554684 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.570684 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.580419 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.589735 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.600930 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.612872 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.625403 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: E0317 15:39:09.627014 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.643548 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.658797 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.691100 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:39:07Z\\\",\\\"message\\\":\\\" 15:39:07.862953 7018 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:39:07.863029 7018 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:39:07.863040 7018 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:39:07.863068 7018 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:39:07.863285 7018 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:39:07.863298 7018 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:39:07.863304 7018 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:39:07.863317 7018 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:39:07.863407 7018 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:39:07.863536 7018 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 15:39:07.863581 7018 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:39:07.863589 7018 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:39:07.863604 7018 factory.go:656] Stopping watch factory\\\\nI0317 15:39:07.863620 7018 ovnkube.go:599] Stopped ovnkube\\\\nI0317 15:39:07.863678 7018 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:39:07.863676 7018 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0317 15\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:39:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.704490 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.736811 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.761543 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.777100 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.790304 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.804937 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.816602 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f49ff1ec-7571-476d-be04-40c50ff291d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2979605c7105c9f289cbfc223356236cea8fb72fec0b0242dea17fb1ad10e950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.840225 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.856047 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.874617 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.885913 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.896139 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:09 crc kubenswrapper[4767]: I0317 15:39:09.908833 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:09Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:10 crc kubenswrapper[4767]: I0317 15:39:10.353702 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:10 crc kubenswrapper[4767]: I0317 15:39:10.353759 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:10 crc kubenswrapper[4767]: I0317 15:39:10.353713 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:10 crc kubenswrapper[4767]: E0317 15:39:10.353902 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:10 crc kubenswrapper[4767]: E0317 15:39:10.354008 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:10 crc kubenswrapper[4767]: E0317 15:39:10.354145 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:11 crc kubenswrapper[4767]: I0317 15:39:11.354034 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:11 crc kubenswrapper[4767]: E0317 15:39:11.354331 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:12 crc kubenswrapper[4767]: I0317 15:39:12.448279 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:12 crc kubenswrapper[4767]: E0317 15:39:12.448446 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:12 crc kubenswrapper[4767]: I0317 15:39:12.448495 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:12 crc kubenswrapper[4767]: I0317 15:39:12.448528 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:12 crc kubenswrapper[4767]: E0317 15:39:12.448673 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:12 crc kubenswrapper[4767]: E0317 15:39:12.448728 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:13 crc kubenswrapper[4767]: I0317 15:39:13.354322 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:13 crc kubenswrapper[4767]: E0317 15:39:13.354439 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:14 crc kubenswrapper[4767]: I0317 15:39:14.354124 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:14 crc kubenswrapper[4767]: I0317 15:39:14.354250 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:14 crc kubenswrapper[4767]: I0317 15:39:14.354250 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:14 crc kubenswrapper[4767]: E0317 15:39:14.355680 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:14 crc kubenswrapper[4767]: E0317 15:39:14.355782 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:14 crc kubenswrapper[4767]: E0317 15:39:14.355888 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:14 crc kubenswrapper[4767]: E0317 15:39:14.628064 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.169989 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.170028 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.170037 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.170054 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.170065 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:15Z","lastTransitionTime":"2026-03-17T15:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:15 crc kubenswrapper[4767]: E0317 15:39:15.185199 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:15Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.189076 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.189218 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.189305 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.189401 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.189461 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:15Z","lastTransitionTime":"2026-03-17T15:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:15 crc kubenswrapper[4767]: E0317 15:39:15.203853 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:15Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.208276 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.208327 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.208339 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.208360 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.208372 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:15Z","lastTransitionTime":"2026-03-17T15:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:15 crc kubenswrapper[4767]: E0317 15:39:15.225384 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:15Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.229713 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.229835 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.229926 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.230047 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.230279 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:15Z","lastTransitionTime":"2026-03-17T15:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:15 crc kubenswrapper[4767]: E0317 15:39:15.246374 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:15Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.250654 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.250689 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.250703 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.250738 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.250752 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:15Z","lastTransitionTime":"2026-03-17T15:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:15 crc kubenswrapper[4767]: E0317 15:39:15.264768 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:15Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:15 crc kubenswrapper[4767]: E0317 15:39:15.265278 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 15:39:15 crc kubenswrapper[4767]: I0317 15:39:15.353976 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:15 crc kubenswrapper[4767]: E0317 15:39:15.354147 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:16 crc kubenswrapper[4767]: I0317 15:39:16.353781 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:16 crc kubenswrapper[4767]: E0317 15:39:16.353977 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:16 crc kubenswrapper[4767]: I0317 15:39:16.353812 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:16 crc kubenswrapper[4767]: I0317 15:39:16.354078 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:16 crc kubenswrapper[4767]: E0317 15:39:16.354404 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:16 crc kubenswrapper[4767]: E0317 15:39:16.354496 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:17 crc kubenswrapper[4767]: I0317 15:39:17.354291 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:17 crc kubenswrapper[4767]: E0317 15:39:17.354514 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:18 crc kubenswrapper[4767]: I0317 15:39:18.353528 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:18 crc kubenswrapper[4767]: I0317 15:39:18.353589 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:18 crc kubenswrapper[4767]: I0317 15:39:18.353646 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:18 crc kubenswrapper[4767]: E0317 15:39:18.353708 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:18 crc kubenswrapper[4767]: E0317 15:39:18.353814 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:18 crc kubenswrapper[4767]: E0317 15:39:18.353926 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.353813 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:19 crc kubenswrapper[4767]: E0317 15:39:19.353948 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.377371 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.393449 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.407077 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.418367 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.429959 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.440705 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.449873 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f49ff1ec-7571-476d-be04-40c50ff291d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2979605c7105c9f289cbfc223356236cea8fb72fec0b0242dea17fb1ad10e950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.462527 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.473001 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.490396 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.499256 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.512961 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.528760 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:39:07Z\\\",\\\"message\\\":\\\" 15:39:07.862953 7018 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:39:07.863029 7018 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:39:07.863040 7018 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:39:07.863068 7018 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:39:07.863285 7018 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:39:07.863298 7018 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:39:07.863304 7018 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:39:07.863317 7018 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:39:07.863407 7018 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:39:07.863536 7018 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 15:39:07.863581 7018 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:39:07.863589 7018 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:39:07.863604 7018 factory.go:656] Stopping watch factory\\\\nI0317 15:39:07.863620 7018 ovnkube.go:599] Stopped ovnkube\\\\nI0317 15:39:07.863678 7018 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:39:07.863676 7018 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0317 15\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:39:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.537370 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.553521 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.575783 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.594636 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: I0317 15:39:19.605504 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:19Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:19 crc kubenswrapper[4767]: E0317 15:39:19.629355 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:39:20 crc kubenswrapper[4767]: I0317 15:39:20.026764 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:20 crc kubenswrapper[4767]: E0317 15:39:20.027033 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:39:20 crc kubenswrapper[4767]: E0317 15:39:20.027118 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs podName:32925427-1529-4e8d-a606-8ccbde1fd877 nodeName:}" failed. No retries permitted until 2026-03-17 15:39:52.027091731 +0000 UTC m=+183.440407778 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs") pod "network-metrics-daemon-t7mfn" (UID: "32925427-1529-4e8d-a606-8ccbde1fd877") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 15:39:20 crc kubenswrapper[4767]: I0317 15:39:20.354095 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:20 crc kubenswrapper[4767]: I0317 15:39:20.354148 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:20 crc kubenswrapper[4767]: I0317 15:39:20.354095 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:20 crc kubenswrapper[4767]: E0317 15:39:20.354299 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:20 crc kubenswrapper[4767]: E0317 15:39:20.354392 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:20 crc kubenswrapper[4767]: E0317 15:39:20.354480 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:21 crc kubenswrapper[4767]: I0317 15:39:21.354142 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:21 crc kubenswrapper[4767]: E0317 15:39:21.354406 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:22 crc kubenswrapper[4767]: I0317 15:39:22.354363 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:22 crc kubenswrapper[4767]: I0317 15:39:22.354382 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:22 crc kubenswrapper[4767]: E0317 15:39:22.354602 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:22 crc kubenswrapper[4767]: E0317 15:39:22.354765 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:22 crc kubenswrapper[4767]: I0317 15:39:22.354384 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:22 crc kubenswrapper[4767]: E0317 15:39:22.354948 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:23 crc kubenswrapper[4767]: I0317 15:39:23.354287 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:23 crc kubenswrapper[4767]: E0317 15:39:23.355337 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:24 crc kubenswrapper[4767]: I0317 15:39:24.354237 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:24 crc kubenswrapper[4767]: I0317 15:39:24.354277 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:24 crc kubenswrapper[4767]: I0317 15:39:24.354311 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:24 crc kubenswrapper[4767]: E0317 15:39:24.354393 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:24 crc kubenswrapper[4767]: E0317 15:39:24.354717 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:24 crc kubenswrapper[4767]: E0317 15:39:24.355113 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:24 crc kubenswrapper[4767]: I0317 15:39:24.355331 4767 scope.go:117] "RemoveContainer" containerID="544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b" Mar 17 15:39:24 crc kubenswrapper[4767]: E0317 15:39:24.355560 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" Mar 17 15:39:24 crc kubenswrapper[4767]: E0317 15:39:24.630186 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.342499 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.342546 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.342560 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.342579 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.342593 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:25Z","lastTransitionTime":"2026-03-17T15:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.354083 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:25 crc kubenswrapper[4767]: E0317 15:39:25.354238 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:25 crc kubenswrapper[4767]: E0317 15:39:25.484034 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:25Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.488257 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.488298 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.488307 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.488322 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.488331 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:25Z","lastTransitionTime":"2026-03-17T15:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:25 crc kubenswrapper[4767]: E0317 15:39:25.502526 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:25Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.505910 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.505947 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.505956 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.505970 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.505980 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:25Z","lastTransitionTime":"2026-03-17T15:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:25 crc kubenswrapper[4767]: E0317 15:39:25.518136 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:25Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.521452 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.521566 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.521641 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.521737 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.521802 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:25Z","lastTransitionTime":"2026-03-17T15:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:25 crc kubenswrapper[4767]: E0317 15:39:25.532584 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:25Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.535815 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.535856 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.535876 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.535893 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:25 crc kubenswrapper[4767]: I0317 15:39:25.535904 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:25Z","lastTransitionTime":"2026-03-17T15:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:25 crc kubenswrapper[4767]: E0317 15:39:25.548069 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bc6e9eb-1e1c-467d-b62d-b25deb35366f\\\",\\\"systemUUID\\\":\\\"8b268e8f-27e3-49d0-9d37-42af1be2e9f2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:25Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:25 crc kubenswrapper[4767]: E0317 15:39:25.548196 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.353855 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.353896 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.353968 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:26 crc kubenswrapper[4767]: E0317 15:39:26.353971 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:26 crc kubenswrapper[4767]: E0317 15:39:26.354098 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:26 crc kubenswrapper[4767]: E0317 15:39:26.354142 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.365086 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.580868 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-v6gnj_f36218dd-ddd3-4269-af32-7fe03113b342/kube-multus/0.log" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.580950 4767 generic.go:334] "Generic (PLEG): container finished" podID="f36218dd-ddd3-4269-af32-7fe03113b342" containerID="6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e" exitCode=1 Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.581011 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-v6gnj" event={"ID":"f36218dd-ddd3-4269-af32-7fe03113b342","Type":"ContainerDied","Data":"6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e"} Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.581908 4767 scope.go:117] "RemoveContainer" containerID="6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.600815 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.619390 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"2026-03-17T15:38:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_01c302bf-95fe-49a2-9221-efa8f5b5b4a9\\\\n2026-03-17T15:38:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_01c302bf-95fe-49a2-9221-efa8f5b5b4a9 to /host/opt/cni/bin/\\\\n2026-03-17T15:38:40Z [verbose] multus-daemon started\\\\n2026-03-17T15:38:40Z [verbose] Readiness Indicator file check\\\\n2026-03-17T15:39:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.633095 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.662533 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:39:07Z\\\",\\\"message\\\":\\\" 15:39:07.862953 7018 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:39:07.863029 7018 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:39:07.863040 7018 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:39:07.863068 7018 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:39:07.863285 7018 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:39:07.863298 7018 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:39:07.863304 7018 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:39:07.863317 7018 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:39:07.863407 7018 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:39:07.863536 7018 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 15:39:07.863581 7018 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:39:07.863589 7018 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:39:07.863604 7018 factory.go:656] Stopping watch factory\\\\nI0317 15:39:07.863620 7018 ovnkube.go:599] Stopped ovnkube\\\\nI0317 15:39:07.863678 7018 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:39:07.863676 7018 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0317 15\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:39:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.681297 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.698948 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b0982b6-fadc-41b0-87ef-a8a0aa48c4c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60814e9521162ab99ae4c6e4913699ef3fdba53dd73a91044129b556b432c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://805aac1aa5eda8c21499dd0ab570c3d30f25303a39357979f5314a0f602a0eb6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:48Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 15:37:18.770697 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 15:37:18.771621 1 observer_polling.go:159] Starting file observer\\\\nI0317 15:37:18.773431 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 15:37:18.774087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 15:37:47.980794 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0317 15:37:48.332734 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 15:37:48.332789 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:18Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:37:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a6d14de833771c3a060a2879802f4f7f37b96b60004b4e5f96374e5fee1428e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c13faf3ff976f76a211bdbb23b12d1165f4905f1a20b7afd81e9250f4b9c792\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://430edad2863573f2d2069c23335b51f5100526f65ab2c9ff799607ed1dbda8ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.721478 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.732597 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.746021 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.771188 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.788797 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.805263 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.821985 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.847493 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.870356 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.883258 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.894704 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f49ff1ec-7571-476d-be04-40c50ff291d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2979605c7105c9f289cbfc223356236cea8fb72fec0b0242dea17fb1ad10e950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.909413 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:26 crc kubenswrapper[4767]: I0317 15:39:26.922985 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:26Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.354013 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:27 crc kubenswrapper[4767]: E0317 15:39:27.354155 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.586808 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-v6gnj_f36218dd-ddd3-4269-af32-7fe03113b342/kube-multus/0.log" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.586884 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-v6gnj" event={"ID":"f36218dd-ddd3-4269-af32-7fe03113b342","Type":"ContainerStarted","Data":"450ee51b49ffbd25e250457c010af40e72c339e723734bcd701ed853117b9041"} Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.600253 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.628087 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:39:07Z\\\",\\\"message\\\":\\\" 15:39:07.862953 7018 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:39:07.863029 7018 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:39:07.863040 7018 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:39:07.863068 7018 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:39:07.863285 7018 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:39:07.863298 7018 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:39:07.863304 7018 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:39:07.863317 7018 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:39:07.863407 7018 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:39:07.863536 7018 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 15:39:07.863581 7018 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:39:07.863589 7018 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:39:07.863604 7018 factory.go:656] Stopping watch factory\\\\nI0317 15:39:07.863620 7018 ovnkube.go:599] Stopped ovnkube\\\\nI0317 15:39:07.863678 7018 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:39:07.863676 7018 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0317 15\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:39:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.640044 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.655124 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b0982b6-fadc-41b0-87ef-a8a0aa48c4c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60814e9521162ab99ae4c6e4913699ef3fdba53dd73a91044129b556b432c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://805aac1aa5eda8c21499dd0ab570c3d30f25303a39357979f5314a0f602a0eb6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:48Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 15:37:18.770697 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 15:37:18.771621 1 observer_polling.go:159] Starting file observer\\\\nI0317 15:37:18.773431 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 15:37:18.774087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 15:37:47.980794 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0317 15:37:48.332734 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 15:37:48.332789 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:18Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:37:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a6d14de833771c3a060a2879802f4f7f37b96b60004b4e5f96374e5fee1428e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c13faf3ff976f76a211bdbb23b12d1165f4905f1a20b7afd81e9250f4b9c792\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://430edad2863573f2d2069c23335b51f5100526f65ab2c9ff799607ed1dbda8ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.667330 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.682646 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.697537 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.722333 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.734309 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.748424 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.757232 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.770054 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.778585 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.788580 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.796996 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f49ff1ec-7571-476d-be04-40c50ff291d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2979605c7105c9f289cbfc223356236cea8fb72fec0b0242dea17fb1ad10e950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.808264 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.817653 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.826731 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:27 crc kubenswrapper[4767]: I0317 15:39:27.837691 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450ee51b49ffbd25e250457c010af40e72c339e723734bcd701ed853117b9041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"2026-03-17T15:38:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_01c302bf-95fe-49a2-9221-efa8f5b5b4a9\\\\n2026-03-17T15:38:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_01c302bf-95fe-49a2-9221-efa8f5b5b4a9 to /host/opt/cni/bin/\\\\n2026-03-17T15:38:40Z [verbose] multus-daemon started\\\\n2026-03-17T15:38:40Z [verbose] Readiness Indicator file check\\\\n2026-03-17T15:39:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:27Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:28 crc kubenswrapper[4767]: I0317 15:39:28.208581 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.208791 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:32.208764939 +0000 UTC m=+223.622080986 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:28 crc kubenswrapper[4767]: I0317 15:39:28.209050 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:28 crc kubenswrapper[4767]: I0317 15:39:28.209164 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.209226 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.209269 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.209468 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:40:32.209396815 +0000 UTC m=+223.622712862 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.209566 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 15:40:32.209553019 +0000 UTC m=+223.622869106 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 15:39:28 crc kubenswrapper[4767]: I0317 15:39:28.310005 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:28 crc kubenswrapper[4767]: I0317 15:39:28.310325 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.310235 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.310520 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.310448 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.310619 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.310634 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.310590 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.310681 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 15:40:32.310666679 +0000 UTC m=+223.723982726 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.310709 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 15:40:32.31069773 +0000 UTC m=+223.724013777 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 15:39:28 crc kubenswrapper[4767]: I0317 15:39:28.354296 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.354419 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:28 crc kubenswrapper[4767]: I0317 15:39:28.354309 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.354593 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:28 crc kubenswrapper[4767]: I0317 15:39:28.354884 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:28 crc kubenswrapper[4767]: E0317 15:39:28.355188 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.354032 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:29 crc kubenswrapper[4767]: E0317 15:39:29.354233 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.368854 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.388548 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.403907 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32925427-1529-4e8d-a606-8ccbde1fd877\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6xc7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t7mfn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.434825 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9779bcaa-2778-48ba-a9bf-ddb4ad9180ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d886e02049095c7edf92ac39df2407e4ed86ca6ebc31d5fd6b37497ae4f5ef12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b46a76dfc16b2697bf2086b5f80b88867bcb51795f8d7e33bf9a98f9490f5acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a97399d7a6335d014b3ce56adc258445b5e3666ab8766bb28bf271cc82d7a3e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14a4f9ba4d27cd1c92ddc178ad22f883bac6c61d8de020ca21fc04890f758514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e120507fadf9ed5d3248a8379d11b4c964723311b4d30eb92e9c5ae352e4429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3f89fa6cc0245b89fce4cbdda423ea99c4dd750f2e8d3cae19f14ed9be274e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d1c21882d59a7992d06507021cc9afda8a4b8a313df276fef078fa332b495a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8837408f01fe5867f550c8a9411cbf3d910f0894b8e38b8bad214785dbab5029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.445755 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f49ff1ec-7571-476d-be04-40c50ff291d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2979605c7105c9f289cbfc223356236cea8fb72fec0b0242dea17fb1ad10e950\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e76cf17a39d372b7bbbc04fc5bb6415f60c88b713d446801ec2b6abca52d241\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.458987 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"786771d9-2ad9-49e9-9c2f-f86e587539a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:44Z\\\",\\\"message\\\":\\\"W0317 15:37:43.875542 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0317 15:37:43.876604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773761863 cert, and key in /tmp/serving-cert-584463218/serving-signer.crt, /tmp/serving-cert-584463218/serving-signer.key\\\\nI0317 15:37:44.310202 1 observer_polling.go:159] Starting file observer\\\\nW0317 15:37:44.316884 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:37:44Z is after 2026-02-23T05:33:16Z\\\\nI0317 15:37:44.317052 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 15:37:44.317672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-584463218/tls.crt::/tmp/serving-cert-584463218/tls.key\\\\\\\"\\\\nF0317 15:37:44.484657 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.472040 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125405fa906c446085d99692f67ec45369ef3738048252cb47ece15fa274e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.489980 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"811a8833-5459-4af6-a1c3-179040eb3c01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf35f4179a04b4abec9b77a921032310041e76d31b64bc9f15e59ec43676175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4522c020c8f65e53f3a395ad1bb491d829d6dbee8daebd401587170925ae1d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82f87546bb869505714d87396022e6387c7039d70f71a29b801cfe2d589a2a47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd7d90cbc9fc24fc6415c745f68d507b685eba270edb21256f4742754c9e9da1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68acfc24c4d6f9a5601eb0053a354996bf7041c4226d3405a0f0e5e76b89215a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747c87568bdb3db1e6503c49405c744f58d2824dc8478a8af1409c8d69ddb60c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4eab053d9713a3036a80732f2c1dd39dda838c89bc96702d5c85492942cbf06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-69rwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nqc7l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.501241 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gdgdx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d82368b-ed78-4142-b1d6-9d86e0aaa810\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b799bf2458f28df7f8ed904190dc9677e07b6c6e5b394b50a3a2a3a0a3d811d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqd4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gdgdx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.516279 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7d39845-ca82-4b08-b5de-e99f0ecbdf18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:37:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac087a6371e7f1d704bf67e050cab18c546ef04adf744a0f2c45a9c353a1a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://679f6f31e5ba44a090ff7af0bf30392438d070b3e4b012666af62557be84905c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8263513c033f5f90b70480feff2e153574104417d3387fb7772d74d6dcf56a46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:36:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.526535 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-v6gnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36218dd-ddd3-4269-af32-7fe03113b342\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450ee51b49ffbd25e250457c010af40e72c339e723734bcd701ed853117b9041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:39:25Z\\\",\\\"message\\\":\\\"2026-03-17T15:38:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_01c302bf-95fe-49a2-9221-efa8f5b5b4a9\\\\n2026-03-17T15:38:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_01c302bf-95fe-49a2-9221-efa8f5b5b4a9 to /host/opt/cni/bin/\\\\n2026-03-17T15:38:40Z [verbose] multus-daemon started\\\\n2026-03-17T15:38:40Z [verbose] Readiness Indicator file check\\\\n2026-03-17T15:39:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wp98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-v6gnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.534285 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7694" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c136104-991b-4c97-ad84-80b82b617b0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bc4dc7f80ed45440c0aec10ef726bfe8ae3feba6121d090bbf7dbe77e71b8dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fn87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7694\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.543627 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a1e15f64097407778a514458aeb5530e4f171fcb33cff940ebc55c0da56e81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.554855 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.565074 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89602a8c1b85ab94626fae54c518fa5edbae0ff7f84cdee68db14e9f3cbc9975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af946388e3d161511b0f321f5f321ebce881855587c78f855258712ca0b134c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.574190 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d680dc75-619e-41ae-90cb-2f58898a33ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a22472c616bf99eab6f893f1c286533cd4339a85b53c630ec0f86396b9fda6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r97fm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lhrtr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.590739 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93753cb4-c3fa-424e-84a2-0a7db20eecfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T15:39:07Z\\\",\\\"message\\\":\\\" 15:39:07.862953 7018 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 15:39:07.863029 7018 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0317 15:39:07.863040 7018 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0317 15:39:07.863068 7018 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0317 15:39:07.863285 7018 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 15:39:07.863298 7018 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 15:39:07.863304 7018 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 15:39:07.863317 7018 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0317 15:39:07.863407 7018 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 15:39:07.863536 7018 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 15:39:07.863581 7018 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 15:39:07.863589 7018 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 15:39:07.863604 7018 factory.go:656] Stopping watch factory\\\\nI0317 15:39:07.863620 7018 ovnkube.go:599] Stopped ovnkube\\\\nI0317 15:39:07.863678 7018 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 15:39:07.863676 7018 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0317 15\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:39:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c4n7r_openshift-ovn-kubernetes(93753cb4-c3fa-424e-84a2-0a7db20eecfa)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T15:38:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T15:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zpdb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c4n7r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.601006 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c277ffee-5af0-4b9e-8afc-206876e483b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0ee4974bb8cdb9a18a04605541a0efab3663b816caaad0e06df2642d619e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8166863c5e835aaa4fb3de5d4385b78dc350ca1b5d12a1c69c4a57d4d7f306d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:38:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-845s5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:38:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bprkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: I0317 15:39:29.613470 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b0982b6-fadc-41b0-87ef-a8a0aa48c4c8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T15:36:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60814e9521162ab99ae4c6e4913699ef3fdba53dd73a91044129b556b432c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://805aac1aa5eda8c21499dd0ab570c3d30f25303a39357979f5314a0f602a0eb6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T15:37:48Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 15:37:18.770697 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 15:37:18.771621 1 observer_polling.go:159] Starting file observer\\\\nI0317 15:37:18.773431 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 15:37:18.774087 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 15:37:47.980794 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0317 15:37:48.332734 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 15:37:48.332789 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T15:37:18Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:37:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a6d14de833771c3a060a2879802f4f7f37b96b60004b4e5f96374e5fee1428e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c13faf3ff976f76a211bdbb23b12d1165f4905f1a20b7afd81e9250f4b9c792\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://430edad2863573f2d2069c23335b51f5100526f65ab2c9ff799607ed1dbda8ed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T15:36:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T15:36:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T15:39:29Z is after 2025-08-24T17:21:41Z" Mar 17 15:39:29 crc kubenswrapper[4767]: E0317 15:39:29.631647 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:39:30 crc kubenswrapper[4767]: I0317 15:39:30.354422 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:30 crc kubenswrapper[4767]: I0317 15:39:30.354448 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:30 crc kubenswrapper[4767]: E0317 15:39:30.354901 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:30 crc kubenswrapper[4767]: I0317 15:39:30.354448 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:30 crc kubenswrapper[4767]: E0317 15:39:30.355019 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:30 crc kubenswrapper[4767]: E0317 15:39:30.355086 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:31 crc kubenswrapper[4767]: I0317 15:39:31.353729 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:31 crc kubenswrapper[4767]: E0317 15:39:31.353912 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:32 crc kubenswrapper[4767]: I0317 15:39:32.353313 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:32 crc kubenswrapper[4767]: I0317 15:39:32.353344 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:32 crc kubenswrapper[4767]: E0317 15:39:32.353436 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:32 crc kubenswrapper[4767]: I0317 15:39:32.353456 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:32 crc kubenswrapper[4767]: E0317 15:39:32.353599 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:32 crc kubenswrapper[4767]: E0317 15:39:32.353629 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:33 crc kubenswrapper[4767]: I0317 15:39:33.353995 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:33 crc kubenswrapper[4767]: E0317 15:39:33.354396 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:34 crc kubenswrapper[4767]: I0317 15:39:34.353772 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:34 crc kubenswrapper[4767]: I0317 15:39:34.353833 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:34 crc kubenswrapper[4767]: I0317 15:39:34.353789 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:34 crc kubenswrapper[4767]: E0317 15:39:34.354027 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:34 crc kubenswrapper[4767]: E0317 15:39:34.354346 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:34 crc kubenswrapper[4767]: E0317 15:39:34.354565 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:34 crc kubenswrapper[4767]: E0317 15:39:34.633487 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.353468 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:35 crc kubenswrapper[4767]: E0317 15:39:35.353971 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.845921 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.845990 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.846013 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.846044 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.846068 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T15:39:35Z","lastTransitionTime":"2026-03-17T15:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.927046 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t"] Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.927612 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.931222 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.931488 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.932064 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.932243 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.993304 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=65.993285495 podStartE2EDuration="1m5.993285495s" podCreationTimestamp="2026-03-17 15:38:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:35.992846204 +0000 UTC m=+167.406162271" watchObservedRunningTime="2026-03-17 15:39:35.993285495 +0000 UTC m=+167.406601542" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.997241 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/556148a9-26ad-411b-976c-17013ad69da3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.997363 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/556148a9-26ad-411b-976c-17013ad69da3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.997454 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/556148a9-26ad-411b-976c-17013ad69da3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.997600 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/556148a9-26ad-411b-976c-17013ad69da3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:35 crc kubenswrapper[4767]: I0317 15:39:35.997699 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/556148a9-26ad-411b-976c-17013ad69da3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.042309 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=65.042281 podStartE2EDuration="1m5.042281s" podCreationTimestamp="2026-03-17 15:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:36.026511476 +0000 UTC m=+167.439827563" watchObservedRunningTime="2026-03-17 15:39:36.042281 +0000 UTC m=+167.455597067" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.063034 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-nqc7l" podStartSLOduration=102.063008411 podStartE2EDuration="1m42.063008411s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:36.062777315 +0000 UTC m=+167.476093382" watchObservedRunningTime="2026-03-17 15:39:36.063008411 +0000 UTC m=+167.476324498" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.074250 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-gdgdx" podStartSLOduration=102.074230288 podStartE2EDuration="1m42.074230288s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:36.073827568 +0000 UTC m=+167.487143625" watchObservedRunningTime="2026-03-17 15:39:36.074230288 +0000 UTC m=+167.487546345" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.087427 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=53.087406356 podStartE2EDuration="53.087406356s" podCreationTimestamp="2026-03-17 15:38:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:36.086696208 +0000 UTC m=+167.500012255" watchObservedRunningTime="2026-03-17 15:39:36.087406356 +0000 UTC m=+167.500722423" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.098236 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/556148a9-26ad-411b-976c-17013ad69da3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.098276 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/556148a9-26ad-411b-976c-17013ad69da3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.098293 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/556148a9-26ad-411b-976c-17013ad69da3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.098311 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/556148a9-26ad-411b-976c-17013ad69da3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.098330 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/556148a9-26ad-411b-976c-17013ad69da3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.098402 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=30.098382157 podStartE2EDuration="30.098382157s" podCreationTimestamp="2026-03-17 15:39:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:36.098038388 +0000 UTC m=+167.511354455" watchObservedRunningTime="2026-03-17 15:39:36.098382157 +0000 UTC m=+167.511698204" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.098575 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/556148a9-26ad-411b-976c-17013ad69da3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.098617 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/556148a9-26ad-411b-976c-17013ad69da3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.099654 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/556148a9-26ad-411b-976c-17013ad69da3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.116480 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/556148a9-26ad-411b-976c-17013ad69da3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.119565 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/556148a9-26ad-411b-976c-17013ad69da3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qsd9t\" (UID: \"556148a9-26ad-411b-976c-17013ad69da3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.146383 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-l7694" podStartSLOduration=102.146359176 podStartE2EDuration="1m42.146359176s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:36.134206554 +0000 UTC m=+167.547522601" watchObservedRunningTime="2026-03-17 15:39:36.146359176 +0000 UTC m=+167.559675223" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.161483 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-v6gnj" podStartSLOduration=102.161465843 podStartE2EDuration="1m42.161465843s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:36.146607622 +0000 UTC m=+167.559923669" watchObservedRunningTime="2026-03-17 15:39:36.161465843 +0000 UTC m=+167.574781890" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.187845 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podStartSLOduration=102.187823588 podStartE2EDuration="1m42.187823588s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:36.187745466 +0000 UTC m=+167.601061523" watchObservedRunningTime="2026-03-17 15:39:36.187823588 +0000 UTC m=+167.601139635" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.225527 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bprkr" podStartSLOduration=101.225508683 podStartE2EDuration="1m41.225508683s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:36.224635081 +0000 UTC m=+167.637951128" watchObservedRunningTime="2026-03-17 15:39:36.225508683 +0000 UTC m=+167.638824730" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.238346 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.257122 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=10.257102442 podStartE2EDuration="10.257102442s" podCreationTimestamp="2026-03-17 15:39:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:36.239394729 +0000 UTC m=+167.652710776" watchObservedRunningTime="2026-03-17 15:39:36.257102442 +0000 UTC m=+167.670418499" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.353543 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.353575 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.353612 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:36 crc kubenswrapper[4767]: E0317 15:39:36.353673 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:36 crc kubenswrapper[4767]: E0317 15:39:36.353776 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:36 crc kubenswrapper[4767]: E0317 15:39:36.353873 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.514241 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.524018 4767 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.619882 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" event={"ID":"556148a9-26ad-411b-976c-17013ad69da3","Type":"ContainerStarted","Data":"48a9269c466360c61b8e41db2aac1f1157cdde86b068241725ada1ba70d925ae"} Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.619969 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" event={"ID":"556148a9-26ad-411b-976c-17013ad69da3","Type":"ContainerStarted","Data":"951e184a78ad5b5ab21cfc2df2b13e5e08361c3859baaa21e5d8286448eaacb3"} Mar 17 15:39:36 crc kubenswrapper[4767]: I0317 15:39:36.638102 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qsd9t" podStartSLOduration=102.63807401 podStartE2EDuration="1m42.63807401s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:36.636420728 +0000 UTC m=+168.049736845" watchObservedRunningTime="2026-03-17 15:39:36.63807401 +0000 UTC m=+168.051390057" Mar 17 15:39:37 crc kubenswrapper[4767]: I0317 15:39:37.354322 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:37 crc kubenswrapper[4767]: E0317 15:39:37.354513 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:37 crc kubenswrapper[4767]: I0317 15:39:37.355626 4767 scope.go:117] "RemoveContainer" containerID="544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b" Mar 17 15:39:37 crc kubenswrapper[4767]: I0317 15:39:37.625418 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovnkube-controller/2.log" Mar 17 15:39:37 crc kubenswrapper[4767]: I0317 15:39:37.627333 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerStarted","Data":"01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f"} Mar 17 15:39:37 crc kubenswrapper[4767]: I0317 15:39:37.627687 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:39:37 crc kubenswrapper[4767]: I0317 15:39:37.654350 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podStartSLOduration=102.654334899 podStartE2EDuration="1m42.654334899s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:37.653136888 +0000 UTC m=+169.066452965" watchObservedRunningTime="2026-03-17 15:39:37.654334899 +0000 UTC m=+169.067650946" Mar 17 15:39:38 crc kubenswrapper[4767]: I0317 15:39:38.353873 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:38 crc kubenswrapper[4767]: E0317 15:39:38.354012 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:38 crc kubenswrapper[4767]: I0317 15:39:38.354281 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:38 crc kubenswrapper[4767]: E0317 15:39:38.354347 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:38 crc kubenswrapper[4767]: I0317 15:39:38.354481 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:38 crc kubenswrapper[4767]: E0317 15:39:38.354539 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:39 crc kubenswrapper[4767]: I0317 15:39:39.344300 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-t7mfn"] Mar 17 15:39:39 crc kubenswrapper[4767]: I0317 15:39:39.344675 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:39 crc kubenswrapper[4767]: E0317 15:39:39.344770 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:39 crc kubenswrapper[4767]: E0317 15:39:39.634716 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:39:40 crc kubenswrapper[4767]: I0317 15:39:40.353644 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:40 crc kubenswrapper[4767]: I0317 15:39:40.353745 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:40 crc kubenswrapper[4767]: E0317 15:39:40.354042 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:40 crc kubenswrapper[4767]: I0317 15:39:40.354082 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:40 crc kubenswrapper[4767]: E0317 15:39:40.354331 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:40 crc kubenswrapper[4767]: E0317 15:39:40.354729 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:41 crc kubenswrapper[4767]: I0317 15:39:41.353725 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:41 crc kubenswrapper[4767]: E0317 15:39:41.353947 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:42 crc kubenswrapper[4767]: I0317 15:39:42.354121 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:42 crc kubenswrapper[4767]: I0317 15:39:42.354253 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:42 crc kubenswrapper[4767]: I0317 15:39:42.354141 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:42 crc kubenswrapper[4767]: E0317 15:39:42.354362 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:42 crc kubenswrapper[4767]: E0317 15:39:42.354544 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:42 crc kubenswrapper[4767]: E0317 15:39:42.354687 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:43 crc kubenswrapper[4767]: I0317 15:39:43.353466 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:43 crc kubenswrapper[4767]: E0317 15:39:43.353660 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t7mfn" podUID="32925427-1529-4e8d-a606-8ccbde1fd877" Mar 17 15:39:44 crc kubenswrapper[4767]: I0317 15:39:44.353252 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:44 crc kubenswrapper[4767]: I0317 15:39:44.353305 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:44 crc kubenswrapper[4767]: I0317 15:39:44.353510 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:44 crc kubenswrapper[4767]: E0317 15:39:44.353501 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 15:39:44 crc kubenswrapper[4767]: E0317 15:39:44.353590 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 15:39:44 crc kubenswrapper[4767]: E0317 15:39:44.353713 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 15:39:45 crc kubenswrapper[4767]: I0317 15:39:45.353832 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:45 crc kubenswrapper[4767]: I0317 15:39:45.357010 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 17 15:39:45 crc kubenswrapper[4767]: I0317 15:39:45.358342 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.016396 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.214707 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rtlwt"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.215007 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.215306 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rbpmk"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.215538 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.216480 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.216844 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.218228 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mdlms"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.218618 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.236572 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.237253 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.237411 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.237560 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.237721 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.237863 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.238009 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.238273 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.238498 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.238658 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.238789 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.238966 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.239882 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.240040 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.241477 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.242199 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5c94c"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.242498 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.242642 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.242717 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.242939 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.243055 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.243904 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.245361 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.265157 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.265612 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rphbj"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.265684 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.266412 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.266810 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.266944 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.267118 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.267248 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.267335 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.267402 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.267516 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.267603 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.268746 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8m5cc"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.269147 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.272229 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.272741 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.273780 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.274086 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.275642 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.278224 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-59hsl"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.278681 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-59hsl" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.279603 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7smc4"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.280114 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.280461 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.280824 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.280854 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.281286 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.283729 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-9cz8q"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.284439 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.284852 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lfctz"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.285436 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.285947 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.286461 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.288584 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.288809 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292145 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292316 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292356 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e0ef009-3626-492a-828e-be5f7285227c-serving-cert\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292377 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07a2b9f1-5291-4204-8631-c1c0f8012bc6-serving-cert\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292394 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6576b15-805f-4a82-ba7b-1e11b5690a91-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bcqcq\" (UID: \"a6576b15-805f-4a82-ba7b-1e11b5690a91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292411 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ece0e760-026e-4de3-bb90-bb5117963f69-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292427 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ece0e760-026e-4de3-bb90-bb5117963f69-audit-policies\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292442 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ece0e760-026e-4de3-bb90-bb5117963f69-encryption-config\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292465 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-config\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292481 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls8gg\" (UniqueName: \"kubernetes.io/projected/deba5665-b026-4d87-a986-b17f02f644d7-kube-api-access-ls8gg\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292498 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4946\" (UniqueName: \"kubernetes.io/projected/a90441f8-30d7-4091-a23f-2522c586f8cf-kube-api-access-p4946\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292515 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbfg6\" (UniqueName: \"kubernetes.io/projected/69925f86-8108-4733-b501-e04fc8cbb320-kube-api-access-vbfg6\") pod \"dns-operator-744455d44c-lfctz\" (UID: \"69925f86-8108-4733-b501-e04fc8cbb320\") " pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292529 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8829d5d-100e-4518-8863-db3ab9c7b30d-serving-cert\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292545 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-policies\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292553 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ldbnt"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292898 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.293269 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.293678 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.293924 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.292559 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f692e30-f8c9-4823-bc82-7bf8abca0659-serving-cert\") pod \"openshift-config-operator-7777fb866f-rphbj\" (UID: \"3f692e30-f8c9-4823-bc82-7bf8abca0659\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294115 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9jxm\" (UniqueName: \"kubernetes.io/projected/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-kube-api-access-m9jxm\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294136 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/69925f86-8108-4733-b501-e04fc8cbb320-metrics-tls\") pod \"dns-operator-744455d44c-lfctz\" (UID: \"69925f86-8108-4733-b501-e04fc8cbb320\") " pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294195 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zp7r\" (UniqueName: \"kubernetes.io/projected/471b6054-3b2a-47c9-8889-942603600da9-kube-api-access-4zp7r\") pod \"downloads-7954f5f757-59hsl\" (UID: \"471b6054-3b2a-47c9-8889-942603600da9\") " pod="openshift-console/downloads-7954f5f757-59hsl" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294235 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294258 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-trusted-ca-bundle\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294279 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ece0e760-026e-4de3-bb90-bb5117963f69-etcd-client\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294298 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ece0e760-026e-4de3-bb90-bb5117963f69-audit-dir\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294322 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/07a2b9f1-5291-4204-8631-c1c0f8012bc6-etcd-service-ca\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294345 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mjfv\" (UniqueName: \"kubernetes.io/projected/3e0ef009-3626-492a-828e-be5f7285227c-kube-api-access-8mjfv\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294352 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294370 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-client-ca\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294395 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xmvv\" (UniqueName: \"kubernetes.io/projected/a6576b15-805f-4a82-ba7b-1e11b5690a91-kube-api-access-8xmvv\") pod \"openshift-apiserver-operator-796bbdcf4f-bcqcq\" (UID: \"a6576b15-805f-4a82-ba7b-1e11b5690a91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294415 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eac52aa6-bd46-4a74-93b1-a1c299eba1c7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c4mkn\" (UID: \"eac52aa6-bd46-4a74-93b1-a1c299eba1c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294441 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a2b9f1-5291-4204-8631-c1c0f8012bc6-config\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294465 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr9xt\" (UniqueName: \"kubernetes.io/projected/3f692e30-f8c9-4823-bc82-7bf8abca0659-kube-api-access-xr9xt\") pod \"openshift-config-operator-7777fb866f-rphbj\" (UID: \"3f692e30-f8c9-4823-bc82-7bf8abca0659\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294486 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8829d5d-100e-4518-8863-db3ab9c7b30d-config\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294509 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-config\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294532 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-serving-cert\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294553 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-config\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294577 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294604 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbf4c\" (UniqueName: \"kubernetes.io/projected/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-kube-api-access-jbf4c\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294625 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-client-ca\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294643 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a90441f8-30d7-4091-a23f-2522c586f8cf-serving-cert\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294664 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3f692e30-f8c9-4823-bc82-7bf8abca0659-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rphbj\" (UID: \"3f692e30-f8c9-4823-bc82-7bf8abca0659\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294707 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294743 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294753 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-console-config\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294780 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294808 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-oauth-serving-cert\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294830 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294854 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-service-ca\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294882 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-dir\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294902 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294928 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294959 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ece0e760-026e-4de3-bb90-bb5117963f69-serving-cert\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.294979 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2130eb2d-c288-473e-931f-68636e3afcea-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7dzwj\" (UID: \"2130eb2d-c288-473e-931f-68636e3afcea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295003 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295032 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-serving-cert\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295535 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2130eb2d-c288-473e-931f-68636e3afcea-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7dzwj\" (UID: \"2130eb2d-c288-473e-931f-68636e3afcea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295560 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295586 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5llj\" (UniqueName: \"kubernetes.io/projected/a8829d5d-100e-4518-8863-db3ab9c7b30d-kube-api-access-j5llj\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295600 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/07a2b9f1-5291-4204-8631-c1c0f8012bc6-etcd-client\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295634 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2t9b\" (UniqueName: \"kubernetes.io/projected/07a2b9f1-5291-4204-8631-c1c0f8012bc6-kube-api-access-n2t9b\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295655 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc5sh\" (UniqueName: \"kubernetes.io/projected/ece0e760-026e-4de3-bb90-bb5117963f69-kube-api-access-zc5sh\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295672 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8829d5d-100e-4518-8863-db3ab9c7b30d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295746 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e0ef009-3626-492a-828e-be5f7285227c-trusted-ca\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295768 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvm4v\" (UniqueName: \"kubernetes.io/projected/eac52aa6-bd46-4a74-93b1-a1c299eba1c7-kube-api-access-hvm4v\") pod \"openshift-controller-manager-operator-756b6f6bc6-c4mkn\" (UID: \"eac52aa6-bd46-4a74-93b1-a1c299eba1c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295783 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/07a2b9f1-5291-4204-8631-c1c0f8012bc6-etcd-ca\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295804 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-images\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295818 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ece0e760-026e-4de3-bb90-bb5117963f69-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295836 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eac52aa6-bd46-4a74-93b1-a1c299eba1c7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c4mkn\" (UID: \"eac52aa6-bd46-4a74-93b1-a1c299eba1c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295850 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8829d5d-100e-4518-8863-db3ab9c7b30d-service-ca-bundle\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295865 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295882 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295896 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6576b15-805f-4a82-ba7b-1e11b5690a91-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bcqcq\" (UID: \"a6576b15-805f-4a82-ba7b-1e11b5690a91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295931 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-oauth-config\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295947 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2130eb2d-c288-473e-931f-68636e3afcea-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7dzwj\" (UID: \"2130eb2d-c288-473e-931f-68636e3afcea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.295964 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mj6f\" (UniqueName: \"kubernetes.io/projected/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-kube-api-access-7mj6f\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.296014 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.296032 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e0ef009-3626-492a-828e-be5f7285227c-config\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.299107 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.299372 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.299488 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.299627 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.299721 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.299828 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.300101 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.300306 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.300756 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.302701 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.308544 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.309265 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.309538 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.309953 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.310387 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.312029 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.312406 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.313060 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.332718 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.461919 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mjfv\" (UniqueName: \"kubernetes.io/projected/3e0ef009-3626-492a-828e-be5f7285227c-kube-api-access-8mjfv\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.462490 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-client-ca\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.462606 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xmvv\" (UniqueName: \"kubernetes.io/projected/a6576b15-805f-4a82-ba7b-1e11b5690a91-kube-api-access-8xmvv\") pod \"openshift-apiserver-operator-796bbdcf4f-bcqcq\" (UID: \"a6576b15-805f-4a82-ba7b-1e11b5690a91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.462796 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eac52aa6-bd46-4a74-93b1-a1c299eba1c7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c4mkn\" (UID: \"eac52aa6-bd46-4a74-93b1-a1c299eba1c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.462919 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a2b9f1-5291-4204-8631-c1c0f8012bc6-config\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.463041 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr9xt\" (UniqueName: \"kubernetes.io/projected/3f692e30-f8c9-4823-bc82-7bf8abca0659-kube-api-access-xr9xt\") pod \"openshift-config-operator-7777fb866f-rphbj\" (UID: \"3f692e30-f8c9-4823-bc82-7bf8abca0659\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.463157 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-config\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.463287 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8829d5d-100e-4518-8863-db3ab9c7b30d-config\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.463414 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.463526 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-serving-cert\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.463660 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-config\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.463744 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a90441f8-30d7-4091-a23f-2522c586f8cf-serving-cert\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.463829 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3f692e30-f8c9-4823-bc82-7bf8abca0659-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rphbj\" (UID: \"3f692e30-f8c9-4823-bc82-7bf8abca0659\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.463954 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbf4c\" (UniqueName: \"kubernetes.io/projected/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-kube-api-access-jbf4c\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.464061 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-client-ca\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.464250 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.464392 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-console-config\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.464508 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.464592 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-oauth-serving-cert\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.464779 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.465021 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-service-ca\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.465217 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.465352 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-dir\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.465475 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.465574 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.465919 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-serving-cert\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.466090 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ece0e760-026e-4de3-bb90-bb5117963f69-serving-cert\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.466341 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2130eb2d-c288-473e-931f-68636e3afcea-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7dzwj\" (UID: \"2130eb2d-c288-473e-931f-68636e3afcea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.466673 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2130eb2d-c288-473e-931f-68636e3afcea-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7dzwj\" (UID: \"2130eb2d-c288-473e-931f-68636e3afcea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.466908 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.467009 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5llj\" (UniqueName: \"kubernetes.io/projected/a8829d5d-100e-4518-8863-db3ab9c7b30d-kube-api-access-j5llj\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.467598 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.467740 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.600005 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.600357 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.600396 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.600556 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.601113 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.602119 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-dir\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.602140 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.602308 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.602407 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.602423 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.602552 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.602586 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/07a2b9f1-5291-4204-8631-c1c0f8012bc6-etcd-client\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.603014 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.603265 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.603379 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.603481 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.603374 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.603809 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.604219 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-ztnfb"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.604264 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.604285 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.604418 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.604418 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.604544 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.604601 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.604667 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.604709 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.605064 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.605074 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.605331 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.605719 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.605862 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.606574 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.607006 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.607292 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.607670 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.607904 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.608048 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.608225 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.608557 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.608641 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.608697 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.608802 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.608869 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.608989 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.609039 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.609129 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.609159 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.608641 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.609331 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.610095 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.608813 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.610874 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.612807 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.613029 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.613048 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2t9b\" (UniqueName: \"kubernetes.io/projected/07a2b9f1-5291-4204-8631-c1c0f8012bc6-kube-api-access-n2t9b\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.613095 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e0ef009-3626-492a-828e-be5f7285227c-trusted-ca\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.613117 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc5sh\" (UniqueName: \"kubernetes.io/projected/ece0e760-026e-4de3-bb90-bb5117963f69-kube-api-access-zc5sh\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.613142 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8829d5d-100e-4518-8863-db3ab9c7b30d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.613203 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvm4v\" (UniqueName: \"kubernetes.io/projected/eac52aa6-bd46-4a74-93b1-a1c299eba1c7-kube-api-access-hvm4v\") pod \"openshift-controller-manager-operator-756b6f6bc6-c4mkn\" (UID: \"eac52aa6-bd46-4a74-93b1-a1c299eba1c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.613290 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/07a2b9f1-5291-4204-8631-c1c0f8012bc6-etcd-ca\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.613319 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.613366 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.613384 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-images\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.614925 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.615847 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.615998 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ece0e760-026e-4de3-bb90-bb5117963f69-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616030 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eac52aa6-bd46-4a74-93b1-a1c299eba1c7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c4mkn\" (UID: \"eac52aa6-bd46-4a74-93b1-a1c299eba1c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616067 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8829d5d-100e-4518-8863-db3ab9c7b30d-service-ca-bundle\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616114 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6576b15-805f-4a82-ba7b-1e11b5690a91-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bcqcq\" (UID: \"a6576b15-805f-4a82-ba7b-1e11b5690a91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616141 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-oauth-config\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616163 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2130eb2d-c288-473e-931f-68636e3afcea-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7dzwj\" (UID: \"2130eb2d-c288-473e-931f-68636e3afcea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616222 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mj6f\" (UniqueName: \"kubernetes.io/projected/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-kube-api-access-7mj6f\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616277 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616311 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e0ef009-3626-492a-828e-be5f7285227c-config\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616335 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07a2b9f1-5291-4204-8631-c1c0f8012bc6-serving-cert\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616371 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616391 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e0ef009-3626-492a-828e-be5f7285227c-serving-cert\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616426 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6576b15-805f-4a82-ba7b-1e11b5690a91-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bcqcq\" (UID: \"a6576b15-805f-4a82-ba7b-1e11b5690a91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616460 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ece0e760-026e-4de3-bb90-bb5117963f69-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616488 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ece0e760-026e-4de3-bb90-bb5117963f69-audit-policies\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616512 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ece0e760-026e-4de3-bb90-bb5117963f69-encryption-config\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616539 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-config\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616564 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls8gg\" (UniqueName: \"kubernetes.io/projected/deba5665-b026-4d87-a986-b17f02f644d7-kube-api-access-ls8gg\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616589 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4946\" (UniqueName: \"kubernetes.io/projected/a90441f8-30d7-4091-a23f-2522c586f8cf-kube-api-access-p4946\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616635 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbfg6\" (UniqueName: \"kubernetes.io/projected/69925f86-8108-4733-b501-e04fc8cbb320-kube-api-access-vbfg6\") pod \"dns-operator-744455d44c-lfctz\" (UID: \"69925f86-8108-4733-b501-e04fc8cbb320\") " pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616657 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8829d5d-100e-4518-8863-db3ab9c7b30d-serving-cert\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616687 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-policies\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616717 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f692e30-f8c9-4823-bc82-7bf8abca0659-serving-cert\") pod \"openshift-config-operator-7777fb866f-rphbj\" (UID: \"3f692e30-f8c9-4823-bc82-7bf8abca0659\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616738 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zp7r\" (UniqueName: \"kubernetes.io/projected/471b6054-3b2a-47c9-8889-942603600da9-kube-api-access-4zp7r\") pod \"downloads-7954f5f757-59hsl\" (UID: \"471b6054-3b2a-47c9-8889-942603600da9\") " pod="openshift-console/downloads-7954f5f757-59hsl" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616765 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9jxm\" (UniqueName: \"kubernetes.io/projected/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-kube-api-access-m9jxm\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616788 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/69925f86-8108-4733-b501-e04fc8cbb320-metrics-tls\") pod \"dns-operator-744455d44c-lfctz\" (UID: \"69925f86-8108-4733-b501-e04fc8cbb320\") " pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616841 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ece0e760-026e-4de3-bb90-bb5117963f69-audit-dir\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616868 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/07a2b9f1-5291-4204-8631-c1c0f8012bc6-etcd-service-ca\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616893 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616914 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-trusted-ca-bundle\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.616933 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ece0e760-026e-4de3-bb90-bb5117963f69-etcd-client\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.618902 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.619287 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.619718 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.619867 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.620700 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.621825 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.622649 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e0ef009-3626-492a-828e-be5f7285227c-trusted-ca\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.624945 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/07a2b9f1-5291-4204-8631-c1c0f8012bc6-etcd-ca\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.626045 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-console-config\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.628468 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-images\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.628813 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.629060 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.629356 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.629941 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.630237 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cb6sk"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.630483 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-cwz8x"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.630849 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vwszx"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.631095 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.631352 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.631663 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.631835 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.631969 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.632087 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.632103 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.632170 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.632214 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.632922 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ece0e760-026e-4de3-bb90-bb5117963f69-serving-cert\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.635314 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2130eb2d-c288-473e-931f-68636e3afcea-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7dzwj\" (UID: \"2130eb2d-c288-473e-931f-68636e3afcea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.635972 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ece0e760-026e-4de3-bb90-bb5117963f69-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.637516 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eac52aa6-bd46-4a74-93b1-a1c299eba1c7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c4mkn\" (UID: \"eac52aa6-bd46-4a74-93b1-a1c299eba1c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.638364 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.639197 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-serving-cert\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.639859 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ece0e760-026e-4de3-bb90-bb5117963f69-audit-policies\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.640249 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.641587 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-client-ca\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.641996 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.642333 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ece0e760-026e-4de3-bb90-bb5117963f69-encryption-config\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.638562 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-client-ca\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.643048 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bk7p5"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.662349 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2130eb2d-c288-473e-931f-68636e3afcea-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7dzwj\" (UID: \"2130eb2d-c288-473e-931f-68636e3afcea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.643420 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.662754 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.662766 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.663640 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.664152 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.664568 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.665164 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.665308 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.665683 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.667050 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.667902 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e0ef009-3626-492a-828e-be5f7285227c-config\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.668320 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-config\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.678991 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8829d5d-100e-4518-8863-db3ab9c7b30d-config\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.679651 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-oauth-serving-cert\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.682159 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3f692e30-f8c9-4823-bc82-7bf8abca0659-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rphbj\" (UID: \"3f692e30-f8c9-4823-bc82-7bf8abca0659\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.684935 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6576b15-805f-4a82-ba7b-1e11b5690a91-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bcqcq\" (UID: \"a6576b15-805f-4a82-ba7b-1e11b5690a91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.685570 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mdlms"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.707379 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.685760 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.685791 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8829d5d-100e-4518-8863-db3ab9c7b30d-service-ca-bundle\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.686283 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.686348 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/07a2b9f1-5291-4204-8631-c1c0f8012bc6-etcd-client\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.686334 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.687115 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.685630 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-service-ca\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.643135 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-config\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.693818 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8829d5d-100e-4518-8863-db3ab9c7b30d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.687184 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.687352 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.687408 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.687593 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.687644 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.722724 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/07a2b9f1-5291-4204-8631-c1c0f8012bc6-etcd-service-ca\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.723134 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.723303 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.723641 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ece0e760-026e-4de3-bb90-bb5117963f69-audit-dir\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.725274 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2130eb2d-c288-473e-931f-68636e3afcea-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7dzwj\" (UID: \"2130eb2d-c288-473e-931f-68636e3afcea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.726331 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-trusted-ca-bundle\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.726780 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/69925f86-8108-4733-b501-e04fc8cbb320-metrics-tls\") pod \"dns-operator-744455d44c-lfctz\" (UID: \"69925f86-8108-4733-b501-e04fc8cbb320\") " pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.726874 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.727614 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-config\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.727673 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a2b9f1-5291-4204-8631-c1c0f8012bc6-config\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.727764 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.727822 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-policies\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.727915 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.728255 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.728518 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ece0e760-026e-4de3-bb90-bb5117963f69-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.728805 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8829d5d-100e-4518-8863-db3ab9c7b30d-serving-cert\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.728842 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.728857 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-serving-cert\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.728882 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ece0e760-026e-4de3-bb90-bb5117963f69-etcd-client\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.729358 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.729613 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e0ef009-3626-492a-828e-be5f7285227c-serving-cert\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.729639 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f692e30-f8c9-4823-bc82-7bf8abca0659-serving-cert\") pod \"openshift-config-operator-7777fb866f-rphbj\" (UID: \"3f692e30-f8c9-4823-bc82-7bf8abca0659\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.729726 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a90441f8-30d7-4091-a23f-2522c586f8cf-serving-cert\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.729786 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6576b15-805f-4a82-ba7b-1e11b5690a91-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bcqcq\" (UID: \"a6576b15-805f-4a82-ba7b-1e11b5690a91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.729859 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.729887 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rbpmk"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.729900 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.730052 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.730301 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.730771 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-oauth-config\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.730881 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-sdx96"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.730933 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.731074 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.731115 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.732468 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.733110 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rphbj"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.733137 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rtlwt"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.733150 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.733161 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.733188 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7smc4"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.733198 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9cz8q"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.733208 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.733216 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.733224 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.733272 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-sdx96" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.733964 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.734241 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.734441 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eac52aa6-bd46-4a74-93b1-a1c299eba1c7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c4mkn\" (UID: \"eac52aa6-bd46-4a74-93b1-a1c299eba1c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.734488 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ldbnt"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.735737 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07a2b9f1-5291-4204-8631-c1c0f8012bc6-serving-cert\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.735802 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8m5cc"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.736525 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.737966 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5c94c"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.738573 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.739272 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-59hsl"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.750560 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.751567 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.752652 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lfctz"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.753898 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.754939 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.755949 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.757649 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cb6sk"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.759886 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-cwz8x"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.760897 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.761987 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.763126 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bdkdk"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.763929 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.764517 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-cspmp"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.765443 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.765871 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-frl2d"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.767283 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.767316 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.768614 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.769713 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.770845 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bk7p5"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.771379 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5llj\" (UniqueName: \"kubernetes.io/projected/a8829d5d-100e-4518-8863-db3ab9c7b30d-kube-api-access-j5llj\") pod \"authentication-operator-69f744f599-mdlms\" (UID: \"a8829d5d-100e-4518-8863-db3ab9c7b30d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.771849 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.772954 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vwszx"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.774494 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.775746 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cspmp"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.776857 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.777851 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-sdx96"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.779119 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-frl2d"] Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.791622 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2t9b\" (UniqueName: \"kubernetes.io/projected/07a2b9f1-5291-4204-8631-c1c0f8012bc6-kube-api-access-n2t9b\") pod \"etcd-operator-b45778765-5c94c\" (UID: \"07a2b9f1-5291-4204-8631-c1c0f8012bc6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.812773 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr9xt\" (UniqueName: \"kubernetes.io/projected/3f692e30-f8c9-4823-bc82-7bf8abca0659-kube-api-access-xr9xt\") pod \"openshift-config-operator-7777fb866f-rphbj\" (UID: \"3f692e30-f8c9-4823-bc82-7bf8abca0659\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.818415 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.835066 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mjfv\" (UniqueName: \"kubernetes.io/projected/3e0ef009-3626-492a-828e-be5f7285227c-kube-api-access-8mjfv\") pod \"console-operator-58897d9998-rbpmk\" (UID: \"3e0ef009-3626-492a-828e-be5f7285227c\") " pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.854938 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xmvv\" (UniqueName: \"kubernetes.io/projected/a6576b15-805f-4a82-ba7b-1e11b5690a91-kube-api-access-8xmvv\") pod \"openshift-apiserver-operator-796bbdcf4f-bcqcq\" (UID: \"a6576b15-805f-4a82-ba7b-1e11b5690a91\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.873662 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc5sh\" (UniqueName: \"kubernetes.io/projected/ece0e760-026e-4de3-bb90-bb5117963f69-kube-api-access-zc5sh\") pod \"apiserver-7bbb656c7d-pv6ds\" (UID: \"ece0e760-026e-4de3-bb90-bb5117963f69\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.891956 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvm4v\" (UniqueName: \"kubernetes.io/projected/eac52aa6-bd46-4a74-93b1-a1c299eba1c7-kube-api-access-hvm4v\") pod \"openshift-controller-manager-operator-756b6f6bc6-c4mkn\" (UID: \"eac52aa6-bd46-4a74-93b1-a1c299eba1c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.894587 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.898531 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.920429 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.938750 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.959733 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 17 15:39:46 crc kubenswrapper[4767]: I0317 15:39:46.961740 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.008592 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.009032 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.018488 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.036412 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.038615 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.042069 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.048486 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.055027 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.058199 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.084434 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.095163 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.098240 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.128048 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.139662 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.160260 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.468481 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.469220 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.472815 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.473001 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.473139 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.473160 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.473408 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.475754 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.476619 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.477638 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.477913 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.478223 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.483032 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.500690 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.501212 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbf4c\" (UniqueName: \"kubernetes.io/projected/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-kube-api-access-jbf4c\") pod \"controller-manager-879f6c89f-rtlwt\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.522415 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.540534 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.822027 4767 request.go:700] Waited for 1.134445787s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/configmaps?fieldSelector=metadata.name%3Dservice-ca-operator-config&limit=500&resourceVersion=0 Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.822661 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bbbc3925-9201-47fa-a29f-e6b271de61fd-proxy-tls\") pod \"machine-config-controller-84d6567774-2cwjx\" (UID: \"bbbc3925-9201-47fa-a29f-e6b271de61fd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.822761 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fwv2\" (UniqueName: \"kubernetes.io/projected/bbbc3925-9201-47fa-a29f-e6b271de61fd-kube-api-access-2fwv2\") pod \"machine-config-controller-84d6567774-2cwjx\" (UID: \"bbbc3925-9201-47fa-a29f-e6b271de61fd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.822821 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-bound-sa-token\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.822879 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3398275a-5195-45d6-9444-9e4d9825770f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.822906 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.823364 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/26899a2e-90e1-49e1-a86b-bfa19c62c195-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jhtfz\" (UID: \"26899a2e-90e1-49e1-a86b-bfa19c62c195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.823659 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-certificates\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.823677 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njbsf\" (UniqueName: \"kubernetes.io/projected/3398275a-5195-45d6-9444-9e4d9825770f-kube-api-access-njbsf\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.823800 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bbbc3925-9201-47fa-a29f-e6b271de61fd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2cwjx\" (UID: \"bbbc3925-9201-47fa-a29f-e6b271de61fd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.823834 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlvw7\" (UniqueName: \"kubernetes.io/projected/26899a2e-90e1-49e1-a86b-bfa19c62c195-kube-api-access-tlvw7\") pod \"control-plane-machine-set-operator-78cbb6b69f-jhtfz\" (UID: \"26899a2e-90e1-49e1-a86b-bfa19c62c195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.823851 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.823850 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3398275a-5195-45d6-9444-9e4d9825770f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.823988 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzmxx\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-kube-api-access-pzmxx\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.824084 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-tls\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.824104 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.824149 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.824166 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3398275a-5195-45d6-9444-9e4d9825770f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.824230 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-trusted-ca\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: E0317 15:39:47.824572 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:48.324561889 +0000 UTC m=+179.737877936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.827842 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.828219 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.828280 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.830732 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936028 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936135 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896af46-1936-4488-9def-e860440a98bf-config\") pod \"service-ca-operator-777779d784-2kkt6\" (UID: \"b896af46-1936-4488-9def-e860440a98bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936171 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3398275a-5195-45d6-9444-9e4d9825770f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936205 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bbbc3925-9201-47fa-a29f-e6b271de61fd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2cwjx\" (UID: \"bbbc3925-9201-47fa-a29f-e6b271de61fd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936220 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98q9s\" (UniqueName: \"kubernetes.io/projected/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-kube-api-access-98q9s\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936239 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqksn\" (UniqueName: \"kubernetes.io/projected/d3064e2c-7080-4395-9a85-8d8794d588e0-kube-api-access-kqksn\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936257 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/75f035cd-32d9-44a2-8d8f-3d5f40742b67-profile-collector-cert\") pod \"catalog-operator-68c6474976-w7cg4\" (UID: \"75f035cd-32d9-44a2-8d8f-3d5f40742b67\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936280 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-tls\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936297 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b896af46-1936-4488-9def-e860440a98bf-serving-cert\") pod \"service-ca-operator-777779d784-2kkt6\" (UID: \"b896af46-1936-4488-9def-e860440a98bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936321 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8d6v\" (UniqueName: \"kubernetes.io/projected/2047989e-3a86-4a4e-8b4a-c8d29fc17e34-kube-api-access-m8d6v\") pod \"multus-admission-controller-857f4d67dd-cwz8x\" (UID: \"2047989e-3a86-4a4e-8b4a-c8d29fc17e34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936342 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936359 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-trusted-ca\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936378 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-config-volume\") pod \"collect-profiles-29562690-fnx5f\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936397 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a5cf3c29-3814-4d17-876a-9ca3e44a9400-default-certificate\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936420 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-webhook-cert\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936441 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/429dd8b3-15c5-4ed5-a81f-1c0f14c1434c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qdjwq\" (UID: \"429dd8b3-15c5-4ed5-a81f-1c0f14c1434c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936456 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dcf5766-ea05-4364-bd69-c51987442223-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-plft5\" (UID: \"2dcf5766-ea05-4364-bd69-c51987442223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936479 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m59wc\" (UniqueName: \"kubernetes.io/projected/b896af46-1936-4488-9def-e860440a98bf-kube-api-access-m59wc\") pod \"service-ca-operator-777779d784-2kkt6\" (UID: \"b896af46-1936-4488-9def-e860440a98bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936506 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5ftw\" (UniqueName: \"kubernetes.io/projected/3b01694a-f9a3-4f54-ae85-7bb50a00f5ca-kube-api-access-k5ftw\") pod \"cluster-samples-operator-665b6dd947-hqjqr\" (UID: \"3b01694a-f9a3-4f54-ae85-7bb50a00f5ca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936520 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55851106-62aa-48a2-bd76-e0049671b476-trusted-ca\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936535 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jfln\" (UniqueName: \"kubernetes.io/projected/02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509-kube-api-access-5jfln\") pod \"ingress-canary-sdx96\" (UID: \"02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509\") " pod="openshift-ingress-canary/ingress-canary-sdx96" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936554 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5cf3c29-3814-4d17-876a-9ca3e44a9400-metrics-certs\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936573 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fwv2\" (UniqueName: \"kubernetes.io/projected/bbbc3925-9201-47fa-a29f-e6b271de61fd-kube-api-access-2fwv2\") pod \"machine-config-controller-84d6567774-2cwjx\" (UID: \"bbbc3925-9201-47fa-a29f-e6b271de61fd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936589 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936603 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7756\" (UniqueName: \"kubernetes.io/projected/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-kube-api-access-r7756\") pod \"collect-profiles-29562690-fnx5f\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936621 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-tmpfs\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936638 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5cf3c29-3814-4d17-876a-9ca3e44a9400-service-ca-bundle\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936653 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/49254887-6adf-47f8-ab3b-8c8890281ee6-metrics-tls\") pod \"dns-default-cspmp\" (UID: \"49254887-6adf-47f8-ab3b-8c8890281ee6\") " pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936667 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkrw2\" (UniqueName: \"kubernetes.io/projected/201932aa-c539-434b-992e-92f889e52de3-kube-api-access-zkrw2\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936681 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/429dd8b3-15c5-4ed5-a81f-1c0f14c1434c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qdjwq\" (UID: \"429dd8b3-15c5-4ed5-a81f-1c0f14c1434c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936699 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-secret-volume\") pod \"collect-profiles-29562690-fnx5f\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936715 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4c4aa56e-7fca-4836-b8ab-581ba3be753c-signing-cabundle\") pod \"service-ca-9c57cc56f-vwszx\" (UID: \"4c4aa56e-7fca-4836-b8ab-581ba3be753c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936729 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-certs\") pod \"machine-config-server-bdkdk\" (UID: \"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe\") " pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936742 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a5cf3c29-3814-4d17-876a-9ca3e44a9400-stats-auth\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936767 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936789 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/68f6ff1d-a119-4faf-89aa-fd9afc2d93f8-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lzjm6\" (UID: \"68f6ff1d-a119-4faf-89aa-fd9afc2d93f8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936806 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-registration-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936823 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmj8n\" (UniqueName: \"kubernetes.io/projected/49254887-6adf-47f8-ab3b-8c8890281ee6-kube-api-access-kmj8n\") pod \"dns-default-cspmp\" (UID: \"49254887-6adf-47f8-ab3b-8c8890281ee6\") " pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936840 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1c88a6ee-cd24-4d85-8a89-d830e5baa434-audit-dir\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936860 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/55851106-62aa-48a2-bd76-e0049671b476-metrics-tls\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936885 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-mountpoint-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936901 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njbsf\" (UniqueName: \"kubernetes.io/projected/3398275a-5195-45d6-9444-9e4d9825770f-kube-api-access-njbsf\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936916 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a333e040-ed72-4d18-95b1-14b57a92c895-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6r96q\" (UID: \"a333e040-ed72-4d18-95b1-14b57a92c895\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936931 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/429dd8b3-15c5-4ed5-a81f-1c0f14c1434c-config\") pod \"kube-controller-manager-operator-78b949d7b-qdjwq\" (UID: \"429dd8b3-15c5-4ed5-a81f-1c0f14c1434c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936945 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjskw\" (UniqueName: \"kubernetes.io/projected/68f6ff1d-a119-4faf-89aa-fd9afc2d93f8-kube-api-access-sjskw\") pod \"package-server-manager-789f6589d5-lzjm6\" (UID: \"68f6ff1d-a119-4faf-89aa-fd9afc2d93f8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936960 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg242\" (UniqueName: \"kubernetes.io/projected/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-kube-api-access-zg242\") pod \"machine-config-server-bdkdk\" (UID: \"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe\") " pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.936985 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6rxd\" (UniqueName: \"kubernetes.io/projected/9b910672-1706-4ce5-8871-e006b25c513d-kube-api-access-p6rxd\") pod \"kube-storage-version-migrator-operator-b67b599dd-dq9tt\" (UID: \"9b910672-1706-4ce5-8871-e006b25c513d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937002 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-apiservice-cert\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937019 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-proxy-tls\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937032 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzpc8\" (UniqueName: \"kubernetes.io/projected/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-kube-api-access-fzpc8\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937053 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-images\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937069 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cb6sk\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937083 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-config\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937097 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvw9j\" (UniqueName: \"kubernetes.io/projected/246c8a34-7be4-4e94-ab60-c4e471a762b1-kube-api-access-pvw9j\") pod \"migrator-59844c95c7-pbr4w\" (UID: \"246c8a34-7be4-4e94-ab60-c4e471a762b1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937114 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlvw7\" (UniqueName: \"kubernetes.io/projected/26899a2e-90e1-49e1-a86b-bfa19c62c195-kube-api-access-tlvw7\") pod \"control-plane-machine-set-operator-78cbb6b69f-jhtfz\" (UID: \"26899a2e-90e1-49e1-a86b-bfa19c62c195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937130 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2047989e-3a86-4a4e-8b4a-c8d29fc17e34-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-cwz8x\" (UID: \"2047989e-3a86-4a4e-8b4a-c8d29fc17e34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937146 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzmxx\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-kube-api-access-pzmxx\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937161 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cb6sk\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937194 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937210 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdkx2\" (UniqueName: \"kubernetes.io/projected/1c88a6ee-cd24-4d85-8a89-d830e5baa434-kube-api-access-mdkx2\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937237 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3398275a-5195-45d6-9444-9e4d9825770f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937257 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwb2x\" (UniqueName: \"kubernetes.io/projected/4c4aa56e-7fca-4836-b8ab-581ba3be753c-kube-api-access-vwb2x\") pod \"service-ca-9c57cc56f-vwszx\" (UID: \"4c4aa56e-7fca-4836-b8ab-581ba3be753c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937272 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9x85\" (UniqueName: \"kubernetes.io/projected/acdeabaa-85fb-4d85-bad6-89fc1be0125a-kube-api-access-g9x85\") pod \"marketplace-operator-79b997595-cb6sk\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937289 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c88a6ee-cd24-4d85-8a89-d830e5baa434-serving-cert\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937305 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b910672-1706-4ce5-8871-e006b25c513d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dq9tt\" (UID: \"9b910672-1706-4ce5-8871-e006b25c513d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937322 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49254887-6adf-47f8-ab3b-8c8890281ee6-config-volume\") pod \"dns-default-cspmp\" (UID: \"49254887-6adf-47f8-ab3b-8c8890281ee6\") " pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937337 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn56j\" (UniqueName: \"kubernetes.io/projected/a333e040-ed72-4d18-95b1-14b57a92c895-kube-api-access-cn56j\") pod \"olm-operator-6b444d44fb-6r96q\" (UID: \"a333e040-ed72-4d18-95b1-14b57a92c895\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937353 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bbbc3925-9201-47fa-a29f-e6b271de61fd-proxy-tls\") pod \"machine-config-controller-84d6567774-2cwjx\" (UID: \"bbbc3925-9201-47fa-a29f-e6b271de61fd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937370 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d3064e2c-7080-4395-9a85-8d8794d588e0-machine-approver-tls\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937387 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-audit\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937403 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2dcf5766-ea05-4364-bd69-c51987442223-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-plft5\" (UID: \"2dcf5766-ea05-4364-bd69-c51987442223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937418 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3b01694a-f9a3-4f54-ae85-7bb50a00f5ca-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hqjqr\" (UID: \"3b01694a-f9a3-4f54-ae85-7bb50a00f5ca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937438 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f98f\" (UniqueName: \"kubernetes.io/projected/75f035cd-32d9-44a2-8d8f-3d5f40742b67-kube-api-access-2f98f\") pod \"catalog-operator-68c6474976-w7cg4\" (UID: \"75f035cd-32d9-44a2-8d8f-3d5f40742b67\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937477 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1c88a6ee-cd24-4d85-8a89-d830e5baa434-encryption-config\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937492 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3064e2c-7080-4395-9a85-8d8794d588e0-config\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937505 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509-cert\") pod \"ingress-canary-sdx96\" (UID: \"02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509\") " pod="openshift-ingress-canary/ingress-canary-sdx96" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937519 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4c4aa56e-7fca-4836-b8ab-581ba3be753c-signing-key\") pod \"service-ca-9c57cc56f-vwszx\" (UID: \"4c4aa56e-7fca-4836-b8ab-581ba3be753c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937535 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-etcd-serving-ca\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937551 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-bound-sa-token\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937567 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-plugins-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937580 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dcf5766-ea05-4364-bd69-c51987442223-config\") pod \"kube-apiserver-operator-766d6c64bb-plft5\" (UID: \"2dcf5766-ea05-4364-bd69-c51987442223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937594 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-socket-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937610 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3398275a-5195-45d6-9444-9e4d9825770f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937631 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1c88a6ee-cd24-4d85-8a89-d830e5baa434-node-pullsecrets\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937653 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-csi-data-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937667 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/55851106-62aa-48a2-bd76-e0049671b476-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937682 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3064e2c-7080-4395-9a85-8d8794d588e0-auth-proxy-config\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937699 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/26899a2e-90e1-49e1-a86b-bfa19c62c195-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jhtfz\" (UID: \"26899a2e-90e1-49e1-a86b-bfa19c62c195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937714 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1c88a6ee-cd24-4d85-8a89-d830e5baa434-etcd-client\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937731 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-node-bootstrap-token\") pod \"machine-config-server-bdkdk\" (UID: \"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe\") " pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937748 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gth44\" (UniqueName: \"kubernetes.io/projected/55851106-62aa-48a2-bd76-e0049671b476-kube-api-access-gth44\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937769 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-certificates\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937787 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/75f035cd-32d9-44a2-8d8f-3d5f40742b67-srv-cert\") pod \"catalog-operator-68c6474976-w7cg4\" (UID: \"75f035cd-32d9-44a2-8d8f-3d5f40742b67\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937828 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsqft\" (UniqueName: \"kubernetes.io/projected/a5cf3c29-3814-4d17-876a-9ca3e44a9400-kube-api-access-xsqft\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937847 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-image-import-ca\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937880 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a333e040-ed72-4d18-95b1-14b57a92c895-srv-cert\") pod \"olm-operator-6b444d44fb-6r96q\" (UID: \"a333e040-ed72-4d18-95b1-14b57a92c895\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.937895 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b910672-1706-4ce5-8871-e006b25c513d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dq9tt\" (UID: \"9b910672-1706-4ce5-8871-e006b25c513d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.939692 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bbbc3925-9201-47fa-a29f-e6b271de61fd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2cwjx\" (UID: \"bbbc3925-9201-47fa-a29f-e6b271de61fd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" Mar 17 15:39:47 crc kubenswrapper[4767]: E0317 15:39:47.939811 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:48.439793861 +0000 UTC m=+179.853109908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.942997 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.946550 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3398275a-5195-45d6-9444-9e4d9825770f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.948717 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.957929 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/26899a2e-90e1-49e1-a86b-bfa19c62c195-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jhtfz\" (UID: \"26899a2e-90e1-49e1-a86b-bfa19c62c195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.958936 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-trusted-ca\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.959377 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3398275a-5195-45d6-9444-9e4d9825770f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.965033 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.965625 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.969019 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.974344 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.974664 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.974842 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.975004 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.975132 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.975215 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.975419 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.975534 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.978219 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bbbc3925-9201-47fa-a29f-e6b271de61fd-proxy-tls\") pod \"machine-config-controller-84d6567774-2cwjx\" (UID: \"bbbc3925-9201-47fa-a29f-e6b271de61fd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.980928 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.981431 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-tls\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.984663 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.990458 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zp7r\" (UniqueName: \"kubernetes.io/projected/471b6054-3b2a-47c9-8889-942603600da9-kube-api-access-4zp7r\") pod \"downloads-7954f5f757-59hsl\" (UID: \"471b6054-3b2a-47c9-8889-942603600da9\") " pod="openshift-console/downloads-7954f5f757-59hsl" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.992481 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbfg6\" (UniqueName: \"kubernetes.io/projected/69925f86-8108-4733-b501-e04fc8cbb320-kube-api-access-vbfg6\") pod \"dns-operator-744455d44c-lfctz\" (UID: \"69925f86-8108-4733-b501-e04fc8cbb320\") " pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.992987 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9jxm\" (UniqueName: \"kubernetes.io/projected/d333e67f-6e87-4ce6-8d3f-4ff16caf2159-kube-api-access-m9jxm\") pod \"machine-api-operator-5694c8668f-7smc4\" (UID: \"d333e67f-6e87-4ce6-8d3f-4ff16caf2159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.993009 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4946\" (UniqueName: \"kubernetes.io/projected/a90441f8-30d7-4091-a23f-2522c586f8cf-kube-api-access-p4946\") pod \"route-controller-manager-6576b87f9c-tz4lw\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:47 crc kubenswrapper[4767]: I0317 15:39:47.993254 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls8gg\" (UniqueName: \"kubernetes.io/projected/deba5665-b026-4d87-a986-b17f02f644d7-kube-api-access-ls8gg\") pod \"console-f9d7485db-9cz8q\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.001123 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-certificates\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.004336 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.007058 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rbpmk"] Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.010522 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.020060 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.021817 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mj6f\" (UniqueName: \"kubernetes.io/projected/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-kube-api-access-7mj6f\") pod \"oauth-openshift-558db77b4-8m5cc\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.039573 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042541 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/75f035cd-32d9-44a2-8d8f-3d5f40742b67-srv-cert\") pod \"catalog-operator-68c6474976-w7cg4\" (UID: \"75f035cd-32d9-44a2-8d8f-3d5f40742b67\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042583 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-node-bootstrap-token\") pod \"machine-config-server-bdkdk\" (UID: \"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe\") " pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042616 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gth44\" (UniqueName: \"kubernetes.io/projected/55851106-62aa-48a2-bd76-e0049671b476-kube-api-access-gth44\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042646 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsqft\" (UniqueName: \"kubernetes.io/projected/a5cf3c29-3814-4d17-876a-9ca3e44a9400-kube-api-access-xsqft\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042671 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-image-import-ca\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042693 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a333e040-ed72-4d18-95b1-14b57a92c895-srv-cert\") pod \"olm-operator-6b444d44fb-6r96q\" (UID: \"a333e040-ed72-4d18-95b1-14b57a92c895\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042715 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b910672-1706-4ce5-8871-e006b25c513d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dq9tt\" (UID: \"9b910672-1706-4ce5-8871-e006b25c513d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042739 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896af46-1936-4488-9def-e860440a98bf-config\") pod \"service-ca-operator-777779d784-2kkt6\" (UID: \"b896af46-1936-4488-9def-e860440a98bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042779 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98q9s\" (UniqueName: \"kubernetes.io/projected/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-kube-api-access-98q9s\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042816 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/75f035cd-32d9-44a2-8d8f-3d5f40742b67-profile-collector-cert\") pod \"catalog-operator-68c6474976-w7cg4\" (UID: \"75f035cd-32d9-44a2-8d8f-3d5f40742b67\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042837 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqksn\" (UniqueName: \"kubernetes.io/projected/d3064e2c-7080-4395-9a85-8d8794d588e0-kube-api-access-kqksn\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042873 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b896af46-1936-4488-9def-e860440a98bf-serving-cert\") pod \"service-ca-operator-777779d784-2kkt6\" (UID: \"b896af46-1936-4488-9def-e860440a98bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042904 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042931 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8d6v\" (UniqueName: \"kubernetes.io/projected/2047989e-3a86-4a4e-8b4a-c8d29fc17e34-kube-api-access-m8d6v\") pod \"multus-admission-controller-857f4d67dd-cwz8x\" (UID: \"2047989e-3a86-4a4e-8b4a-c8d29fc17e34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042961 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.042996 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-config-volume\") pod \"collect-profiles-29562690-fnx5f\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043017 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a5cf3c29-3814-4d17-876a-9ca3e44a9400-default-certificate\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043045 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-webhook-cert\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043073 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/429dd8b3-15c5-4ed5-a81f-1c0f14c1434c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qdjwq\" (UID: \"429dd8b3-15c5-4ed5-a81f-1c0f14c1434c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043100 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dcf5766-ea05-4364-bd69-c51987442223-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-plft5\" (UID: \"2dcf5766-ea05-4364-bd69-c51987442223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043122 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m59wc\" (UniqueName: \"kubernetes.io/projected/b896af46-1936-4488-9def-e860440a98bf-kube-api-access-m59wc\") pod \"service-ca-operator-777779d784-2kkt6\" (UID: \"b896af46-1936-4488-9def-e860440a98bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043195 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jfln\" (UniqueName: \"kubernetes.io/projected/02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509-kube-api-access-5jfln\") pod \"ingress-canary-sdx96\" (UID: \"02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509\") " pod="openshift-ingress-canary/ingress-canary-sdx96" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043225 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5ftw\" (UniqueName: \"kubernetes.io/projected/3b01694a-f9a3-4f54-ae85-7bb50a00f5ca-kube-api-access-k5ftw\") pod \"cluster-samples-operator-665b6dd947-hqjqr\" (UID: \"3b01694a-f9a3-4f54-ae85-7bb50a00f5ca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043251 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55851106-62aa-48a2-bd76-e0049671b476-trusted-ca\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043275 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5cf3c29-3814-4d17-876a-9ca3e44a9400-metrics-certs\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043316 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043345 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7756\" (UniqueName: \"kubernetes.io/projected/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-kube-api-access-r7756\") pod \"collect-profiles-29562690-fnx5f\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043374 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-tmpfs\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043399 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/49254887-6adf-47f8-ab3b-8c8890281ee6-metrics-tls\") pod \"dns-default-cspmp\" (UID: \"49254887-6adf-47f8-ab3b-8c8890281ee6\") " pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043420 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkrw2\" (UniqueName: \"kubernetes.io/projected/201932aa-c539-434b-992e-92f889e52de3-kube-api-access-zkrw2\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043450 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5cf3c29-3814-4d17-876a-9ca3e44a9400-service-ca-bundle\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043474 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/429dd8b3-15c5-4ed5-a81f-1c0f14c1434c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qdjwq\" (UID: \"429dd8b3-15c5-4ed5-a81f-1c0f14c1434c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043495 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-secret-volume\") pod \"collect-profiles-29562690-fnx5f\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043519 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4c4aa56e-7fca-4836-b8ab-581ba3be753c-signing-cabundle\") pod \"service-ca-9c57cc56f-vwszx\" (UID: \"4c4aa56e-7fca-4836-b8ab-581ba3be753c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043544 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-certs\") pod \"machine-config-server-bdkdk\" (UID: \"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe\") " pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043565 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a5cf3c29-3814-4d17-876a-9ca3e44a9400-stats-auth\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043588 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-registration-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043618 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/68f6ff1d-a119-4faf-89aa-fd9afc2d93f8-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lzjm6\" (UID: \"68f6ff1d-a119-4faf-89aa-fd9afc2d93f8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043652 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmj8n\" (UniqueName: \"kubernetes.io/projected/49254887-6adf-47f8-ab3b-8c8890281ee6-kube-api-access-kmj8n\") pod \"dns-default-cspmp\" (UID: \"49254887-6adf-47f8-ab3b-8c8890281ee6\") " pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043678 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1c88a6ee-cd24-4d85-8a89-d830e5baa434-audit-dir\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043701 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/55851106-62aa-48a2-bd76-e0049671b476-metrics-tls\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043737 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-mountpoint-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043769 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a333e040-ed72-4d18-95b1-14b57a92c895-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6r96q\" (UID: \"a333e040-ed72-4d18-95b1-14b57a92c895\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043794 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/429dd8b3-15c5-4ed5-a81f-1c0f14c1434c-config\") pod \"kube-controller-manager-operator-78b949d7b-qdjwq\" (UID: \"429dd8b3-15c5-4ed5-a81f-1c0f14c1434c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043818 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjskw\" (UniqueName: \"kubernetes.io/projected/68f6ff1d-a119-4faf-89aa-fd9afc2d93f8-kube-api-access-sjskw\") pod \"package-server-manager-789f6589d5-lzjm6\" (UID: \"68f6ff1d-a119-4faf-89aa-fd9afc2d93f8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043842 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg242\" (UniqueName: \"kubernetes.io/projected/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-kube-api-access-zg242\") pod \"machine-config-server-bdkdk\" (UID: \"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe\") " pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043866 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6rxd\" (UniqueName: \"kubernetes.io/projected/9b910672-1706-4ce5-8871-e006b25c513d-kube-api-access-p6rxd\") pod \"kube-storage-version-migrator-operator-b67b599dd-dq9tt\" (UID: \"9b910672-1706-4ce5-8871-e006b25c513d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043897 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-apiservice-cert\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043933 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-proxy-tls\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043957 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzpc8\" (UniqueName: \"kubernetes.io/projected/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-kube-api-access-fzpc8\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.043981 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-config\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044006 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvw9j\" (UniqueName: \"kubernetes.io/projected/246c8a34-7be4-4e94-ab60-c4e471a762b1-kube-api-access-pvw9j\") pod \"migrator-59844c95c7-pbr4w\" (UID: \"246c8a34-7be4-4e94-ab60-c4e471a762b1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044027 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-images\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044053 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cb6sk\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044096 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2047989e-3a86-4a4e-8b4a-c8d29fc17e34-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-cwz8x\" (UID: \"2047989e-3a86-4a4e-8b4a-c8d29fc17e34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044120 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdkx2\" (UniqueName: \"kubernetes.io/projected/1c88a6ee-cd24-4d85-8a89-d830e5baa434-kube-api-access-mdkx2\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044144 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cb6sk\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044227 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c88a6ee-cd24-4d85-8a89-d830e5baa434-serving-cert\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044252 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b910672-1706-4ce5-8871-e006b25c513d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dq9tt\" (UID: \"9b910672-1706-4ce5-8871-e006b25c513d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044275 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwb2x\" (UniqueName: \"kubernetes.io/projected/4c4aa56e-7fca-4836-b8ab-581ba3be753c-kube-api-access-vwb2x\") pod \"service-ca-9c57cc56f-vwszx\" (UID: \"4c4aa56e-7fca-4836-b8ab-581ba3be753c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044295 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9x85\" (UniqueName: \"kubernetes.io/projected/acdeabaa-85fb-4d85-bad6-89fc1be0125a-kube-api-access-g9x85\") pod \"marketplace-operator-79b997595-cb6sk\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044331 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49254887-6adf-47f8-ab3b-8c8890281ee6-config-volume\") pod \"dns-default-cspmp\" (UID: \"49254887-6adf-47f8-ab3b-8c8890281ee6\") " pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044354 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn56j\" (UniqueName: \"kubernetes.io/projected/a333e040-ed72-4d18-95b1-14b57a92c895-kube-api-access-cn56j\") pod \"olm-operator-6b444d44fb-6r96q\" (UID: \"a333e040-ed72-4d18-95b1-14b57a92c895\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044377 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d3064e2c-7080-4395-9a85-8d8794d588e0-machine-approver-tls\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044398 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-audit\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044419 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2dcf5766-ea05-4364-bd69-c51987442223-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-plft5\" (UID: \"2dcf5766-ea05-4364-bd69-c51987442223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044443 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f98f\" (UniqueName: \"kubernetes.io/projected/75f035cd-32d9-44a2-8d8f-3d5f40742b67-kube-api-access-2f98f\") pod \"catalog-operator-68c6474976-w7cg4\" (UID: \"75f035cd-32d9-44a2-8d8f-3d5f40742b67\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044465 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1c88a6ee-cd24-4d85-8a89-d830e5baa434-encryption-config\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044485 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3b01694a-f9a3-4f54-ae85-7bb50a00f5ca-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hqjqr\" (UID: \"3b01694a-f9a3-4f54-ae85-7bb50a00f5ca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044505 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4c4aa56e-7fca-4836-b8ab-581ba3be753c-signing-key\") pod \"service-ca-9c57cc56f-vwszx\" (UID: \"4c4aa56e-7fca-4836-b8ab-581ba3be753c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044530 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3064e2c-7080-4395-9a85-8d8794d588e0-config\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044550 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509-cert\") pod \"ingress-canary-sdx96\" (UID: \"02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509\") " pod="openshift-ingress-canary/ingress-canary-sdx96" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044572 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-etcd-serving-ca\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044598 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-plugins-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044621 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dcf5766-ea05-4364-bd69-c51987442223-config\") pod \"kube-apiserver-operator-766d6c64bb-plft5\" (UID: \"2dcf5766-ea05-4364-bd69-c51987442223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044642 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-socket-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044680 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1c88a6ee-cd24-4d85-8a89-d830e5baa434-node-pullsecrets\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044702 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1c88a6ee-cd24-4d85-8a89-d830e5baa434-etcd-client\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044721 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-csi-data-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044745 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/55851106-62aa-48a2-bd76-e0049671b476-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.044791 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3064e2c-7080-4395-9a85-8d8794d588e0-auth-proxy-config\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:48 crc kubenswrapper[4767]: E0317 15:39:48.044985 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:48.544967885 +0000 UTC m=+179.958284002 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.045728 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896af46-1936-4488-9def-e860440a98bf-config\") pod \"service-ca-operator-777779d784-2kkt6\" (UID: \"b896af46-1936-4488-9def-e860440a98bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.049253 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55851106-62aa-48a2-bd76-e0049671b476-trusted-ca\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.049404 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-images\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.050667 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cb6sk\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.051414 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-mountpoint-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.052385 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/75f035cd-32d9-44a2-8d8f-3d5f40742b67-srv-cert\") pod \"catalog-operator-68c6474976-w7cg4\" (UID: \"75f035cd-32d9-44a2-8d8f-3d5f40742b67\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.052639 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.053093 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-tmpfs\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.053525 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4c4aa56e-7fca-4836-b8ab-581ba3be753c-signing-cabundle\") pod \"service-ca-9c57cc56f-vwszx\" (UID: \"4c4aa56e-7fca-4836-b8ab-581ba3be753c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.053837 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5cf3c29-3814-4d17-876a-9ca3e44a9400-metrics-certs\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.054162 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-registration-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.054780 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-plugins-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.054998 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/429dd8b3-15c5-4ed5-a81f-1c0f14c1434c-config\") pod \"kube-controller-manager-operator-78b949d7b-qdjwq\" (UID: \"429dd8b3-15c5-4ed5-a81f-1c0f14c1434c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.055119 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b910672-1706-4ce5-8871-e006b25c513d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dq9tt\" (UID: \"9b910672-1706-4ce5-8871-e006b25c513d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.055287 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1c88a6ee-cd24-4d85-8a89-d830e5baa434-audit-dir\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.055712 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2047989e-3a86-4a4e-8b4a-c8d29fc17e34-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-cwz8x\" (UID: \"2047989e-3a86-4a4e-8b4a-c8d29fc17e34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.055859 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-socket-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.055886 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1c88a6ee-cd24-4d85-8a89-d830e5baa434-node-pullsecrets\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.056019 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dcf5766-ea05-4364-bd69-c51987442223-config\") pod \"kube-apiserver-operator-766d6c64bb-plft5\" (UID: \"2dcf5766-ea05-4364-bd69-c51987442223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.056089 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/201932aa-c539-434b-992e-92f889e52de3-csi-data-dir\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.056573 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5cf3c29-3814-4d17-876a-9ca3e44a9400-service-ca-bundle\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.056949 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b896af46-1936-4488-9def-e860440a98bf-serving-cert\") pod \"service-ca-operator-777779d784-2kkt6\" (UID: \"b896af46-1936-4488-9def-e860440a98bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.058003 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a5cf3c29-3814-4d17-876a-9ca3e44a9400-stats-auth\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.058025 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.058427 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1c88a6ee-cd24-4d85-8a89-d830e5baa434-encryption-config\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.059260 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3b01694a-f9a3-4f54-ae85-7bb50a00f5ca-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hqjqr\" (UID: \"3b01694a-f9a3-4f54-ae85-7bb50a00f5ca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.059779 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cb6sk\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.060815 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/429dd8b3-15c5-4ed5-a81f-1c0f14c1434c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qdjwq\" (UID: \"429dd8b3-15c5-4ed5-a81f-1c0f14c1434c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.060822 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a5cf3c29-3814-4d17-876a-9ca3e44a9400-default-certificate\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.060968 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b910672-1706-4ce5-8871-e006b25c513d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dq9tt\" (UID: \"9b910672-1706-4ce5-8871-e006b25c513d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.061075 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-secret-volume\") pod \"collect-profiles-29562690-fnx5f\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.061407 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-proxy-tls\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.061827 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/75f035cd-32d9-44a2-8d8f-3d5f40742b67-profile-collector-cert\") pod \"catalog-operator-68c6474976-w7cg4\" (UID: \"75f035cd-32d9-44a2-8d8f-3d5f40742b67\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.062543 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a333e040-ed72-4d18-95b1-14b57a92c895-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6r96q\" (UID: \"a333e040-ed72-4d18-95b1-14b57a92c895\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.062857 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/55851106-62aa-48a2-bd76-e0049671b476-metrics-tls\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.065985 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dcf5766-ea05-4364-bd69-c51987442223-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-plft5\" (UID: \"2dcf5766-ea05-4364-bd69-c51987442223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.071882 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4c4aa56e-7fca-4836-b8ab-581ba3be753c-signing-key\") pod \"service-ca-9c57cc56f-vwszx\" (UID: \"4c4aa56e-7fca-4836-b8ab-581ba3be753c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.189481 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:48 crc kubenswrapper[4767]: E0317 15:39:48.190043 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:48.690027769 +0000 UTC m=+180.103343816 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.232703 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.232985 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.233151 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.233321 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.233441 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.233550 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.234813 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-audit\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.235989 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-config-volume\") pod \"collect-profiles-29562690-fnx5f\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.237101 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.238527 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-config\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.238805 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.245787 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-image-import-ca\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.247059 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.250450 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/68f6ff1d-a119-4faf-89aa-fd9afc2d93f8-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lzjm6\" (UID: \"68f6ff1d-a119-4faf-89aa-fd9afc2d93f8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.250751 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-etcd-serving-ca\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.253640 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1c88a6ee-cd24-4d85-8a89-d830e5baa434-etcd-client\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.257641 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c88a6ee-cd24-4d85-8a89-d830e5baa434-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.258438 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.259320 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c88a6ee-cd24-4d85-8a89-d830e5baa434-serving-cert\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.261386 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj"] Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.263403 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.269797 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-59hsl" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.279637 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.279872 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.280425 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds"] Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.288198 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.292470 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:48 crc kubenswrapper[4767]: E0317 15:39:48.292969 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:48.792949005 +0000 UTC m=+180.206265052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.299508 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.325155 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.344618 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.344758 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d3064e2c-7080-4395-9a85-8d8794d588e0-machine-approver-tls\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.347252 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3064e2c-7080-4395-9a85-8d8794d588e0-config\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.358256 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.366033 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3064e2c-7080-4395-9a85-8d8794d588e0-auth-proxy-config\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:48 crc kubenswrapper[4767]: I0317 15:39:48.407720 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.161629 4767 request.go:700] Waited for 1.115585112s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/serviceaccounts/machine-config-operator/token Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.162290 4767 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.162351 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-certs podName:3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe nodeName:}" failed. No retries permitted until 2026-03-17 15:39:49.662333443 +0000 UTC m=+181.075649490 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-certs") pod "machine-config-server-bdkdk" (UID: "3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe") : failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.163718 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.164268 4767 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.164363 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509-cert podName:02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509 nodeName:}" failed. No retries permitted until 2026-03-17 15:39:49.664340825 +0000 UTC m=+181.077656872 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509-cert") pod "ingress-canary-sdx96" (UID: "02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509") : failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.164575 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:50.164563161 +0000 UTC m=+181.577879208 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.165264 4767 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.165311 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49254887-6adf-47f8-ab3b-8c8890281ee6-metrics-tls podName:49254887-6adf-47f8-ab3b-8c8890281ee6 nodeName:}" failed. No retries permitted until 2026-03-17 15:39:49.665298599 +0000 UTC m=+181.078614726 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/49254887-6adf-47f8-ab3b-8c8890281ee6-metrics-tls") pod "dns-default-cspmp" (UID: "49254887-6adf-47f8-ab3b-8c8890281ee6") : failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.166433 4767 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.166473 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/49254887-6adf-47f8-ab3b-8c8890281ee6-config-volume podName:49254887-6adf-47f8-ab3b-8c8890281ee6 nodeName:}" failed. No retries permitted until 2026-03-17 15:39:49.666462469 +0000 UTC m=+181.079778576 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/49254887-6adf-47f8-ab3b-8c8890281ee6-config-volume") pod "dns-default-cspmp" (UID: "49254887-6adf-47f8-ab3b-8c8890281ee6") : failed to sync configmap cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.172122 4767 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.172207 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-apiservice-cert podName:b836be84-e8b0-4e8d-8bac-cfc922bea7eb nodeName:}" failed. No retries permitted until 2026-03-17 15:39:49.672191386 +0000 UTC m=+181.085507433 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-apiservice-cert") pod "packageserver-d55dfcdfc-8rfjw" (UID: "b836be84-e8b0-4e8d-8bac-cfc922bea7eb") : failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.172240 4767 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.172273 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-webhook-cert podName:b836be84-e8b0-4e8d-8bac-cfc922bea7eb nodeName:}" failed. No retries permitted until 2026-03-17 15:39:49.672267538 +0000 UTC m=+181.085583585 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-webhook-cert") pod "packageserver-d55dfcdfc-8rfjw" (UID: "b836be84-e8b0-4e8d-8bac-cfc922bea7eb") : failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.172491 4767 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.172533 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-node-bootstrap-token podName:3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe nodeName:}" failed. No retries permitted until 2026-03-17 15:39:49.672526254 +0000 UTC m=+181.085842301 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-node-bootstrap-token") pod "machine-config-server-bdkdk" (UID: "3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe") : failed to sync secret cache: timed out waiting for the condition Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.186966 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a333e040-ed72-4d18-95b1-14b57a92c895-srv-cert\") pod \"olm-operator-6b444d44fb-6r96q\" (UID: \"a333e040-ed72-4d18-95b1-14b57a92c895\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.232956 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.233217 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.233277 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.233346 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.238233 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjskw\" (UniqueName: \"kubernetes.io/projected/68f6ff1d-a119-4faf-89aa-fd9afc2d93f8-kube-api-access-sjskw\") pod \"package-server-manager-789f6589d5-lzjm6\" (UID: \"68f6ff1d-a119-4faf-89aa-fd9afc2d93f8\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.240889 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsqft\" (UniqueName: \"kubernetes.io/projected/a5cf3c29-3814-4d17-876a-9ca3e44a9400-kube-api-access-xsqft\") pod \"router-default-5444994796-ztnfb\" (UID: \"a5cf3c29-3814-4d17-876a-9ca3e44a9400\") " pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.241871 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlvw7\" (UniqueName: \"kubernetes.io/projected/26899a2e-90e1-49e1-a86b-bfa19c62c195-kube-api-access-tlvw7\") pod \"control-plane-machine-set-operator-78cbb6b69f-jhtfz\" (UID: \"26899a2e-90e1-49e1-a86b-bfa19c62c195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.244365 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8d6v\" (UniqueName: \"kubernetes.io/projected/2047989e-3a86-4a4e-8b4a-c8d29fc17e34-kube-api-access-m8d6v\") pod \"multus-admission-controller-857f4d67dd-cwz8x\" (UID: \"2047989e-3a86-4a4e-8b4a-c8d29fc17e34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.245014 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.247301 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m59wc\" (UniqueName: \"kubernetes.io/projected/b896af46-1936-4488-9def-e860440a98bf-kube-api-access-m59wc\") pod \"service-ca-operator-777779d784-2kkt6\" (UID: \"b896af46-1936-4488-9def-e860440a98bf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.247593 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.247863 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.247999 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.248126 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.249325 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.249524 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.263314 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" event={"ID":"2130eb2d-c288-473e-931f-68636e3afcea","Type":"ContainerStarted","Data":"de767763723f860c4a5674d1003c2448149a1ccd703ab646a38f56135944d609"} Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.267239 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn"] Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.264301 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.264113 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.265787 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.267965 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq"] Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.268856 4767 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.270033 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.270847 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:49.770823952 +0000 UTC m=+181.184139999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.272417 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mdlms"] Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.273149 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jfln\" (UniqueName: \"kubernetes.io/projected/02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509-kube-api-access-5jfln\") pod \"ingress-canary-sdx96\" (UID: \"02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509\") " pod="openshift-ingress-canary/ingress-canary-sdx96" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.273293 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rphbj"] Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.273318 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" event={"ID":"3e0ef009-3626-492a-828e-be5f7285227c","Type":"ContainerStarted","Data":"0adbf15d1072c1fe23beb1443ff1648323f80fd1ab7345b545ce8b2cdd2060e7"} Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.273347 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" event={"ID":"3e0ef009-3626-492a-828e-be5f7285227c","Type":"ContainerStarted","Data":"1a3830eadcdfa17190d8c02761df6bbe54b5b70eeb32be1db21288b2fa55a930"} Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.273594 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.274060 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.276348 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6rxd\" (UniqueName: \"kubernetes.io/projected/9b910672-1706-4ce5-8871-e006b25c513d-kube-api-access-p6rxd\") pod \"kube-storage-version-migrator-operator-b67b599dd-dq9tt\" (UID: \"9b910672-1706-4ce5-8871-e006b25c513d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.277113 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" event={"ID":"ece0e760-026e-4de3-bb90-bb5117963f69","Type":"ContainerStarted","Data":"b70f18f21c801a34adee033e805e7351cd65b3284f79f6a250874955e34db2e8"} Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.277775 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gth44\" (UniqueName: \"kubernetes.io/projected/55851106-62aa-48a2-bd76-e0049671b476-kube-api-access-gth44\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.278289 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f98f\" (UniqueName: \"kubernetes.io/projected/75f035cd-32d9-44a2-8d8f-3d5f40742b67-kube-api-access-2f98f\") pod \"catalog-operator-68c6474976-w7cg4\" (UID: \"75f035cd-32d9-44a2-8d8f-3d5f40742b67\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.278448 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njbsf\" (UniqueName: \"kubernetes.io/projected/3398275a-5195-45d6-9444-9e4d9825770f-kube-api-access-njbsf\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.279042 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9x85\" (UniqueName: \"kubernetes.io/projected/acdeabaa-85fb-4d85-bad6-89fc1be0125a-kube-api-access-g9x85\") pod \"marketplace-operator-79b997595-cb6sk\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.279547 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdkx2\" (UniqueName: \"kubernetes.io/projected/1c88a6ee-cd24-4d85-8a89-d830e5baa434-kube-api-access-mdkx2\") pod \"apiserver-76f77b778f-bk7p5\" (UID: \"1c88a6ee-cd24-4d85-8a89-d830e5baa434\") " pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.279821 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzmxx\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-kube-api-access-pzmxx\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.281846 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-bound-sa-token\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.283441 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg242\" (UniqueName: \"kubernetes.io/projected/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-kube-api-access-zg242\") pod \"machine-config-server-bdkdk\" (UID: \"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe\") " pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.284090 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/429dd8b3-15c5-4ed5-a81f-1c0f14c1434c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qdjwq\" (UID: \"429dd8b3-15c5-4ed5-a81f-1c0f14c1434c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.284979 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.285071 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.285135 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2dcf5766-ea05-4364-bd69-c51987442223-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-plft5\" (UID: \"2dcf5766-ea05-4364-bd69-c51987442223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.285646 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fwv2\" (UniqueName: \"kubernetes.io/projected/bbbc3925-9201-47fa-a29f-e6b271de61fd-kube-api-access-2fwv2\") pod \"machine-config-controller-84d6567774-2cwjx\" (UID: \"bbbc3925-9201-47fa-a29f-e6b271de61fd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.288766 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvw9j\" (UniqueName: \"kubernetes.io/projected/246c8a34-7be4-4e94-ab60-c4e471a762b1-kube-api-access-pvw9j\") pod \"migrator-59844c95c7-pbr4w\" (UID: \"246c8a34-7be4-4e94-ab60-c4e471a762b1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.289239 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzpc8\" (UniqueName: \"kubernetes.io/projected/99e547df-6c91-4f2b-b8c8-5a9ddcb942b3-kube-api-access-fzpc8\") pod \"machine-config-operator-74547568cd-2v6ms\" (UID: \"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.290736 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5ftw\" (UniqueName: \"kubernetes.io/projected/3b01694a-f9a3-4f54-ae85-7bb50a00f5ca-kube-api-access-k5ftw\") pod \"cluster-samples-operator-665b6dd947-hqjqr\" (UID: \"3b01694a-f9a3-4f54-ae85-7bb50a00f5ca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.295308 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lfctz"] Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.302516 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkrw2\" (UniqueName: \"kubernetes.io/projected/201932aa-c539-434b-992e-92f889e52de3-kube-api-access-zkrw2\") pod \"csi-hostpathplugin-frl2d\" (UID: \"201932aa-c539-434b-992e-92f889e52de3\") " pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.302851 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3398275a-5195-45d6-9444-9e4d9825770f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k4twt\" (UID: \"3398275a-5195-45d6-9444-9e4d9825770f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.304216 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7756\" (UniqueName: \"kubernetes.io/projected/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-kube-api-access-r7756\") pod \"collect-profiles-29562690-fnx5f\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.305025 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98q9s\" (UniqueName: \"kubernetes.io/projected/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-kube-api-access-98q9s\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.305283 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn56j\" (UniqueName: \"kubernetes.io/projected/a333e040-ed72-4d18-95b1-14b57a92c895-kube-api-access-cn56j\") pod \"olm-operator-6b444d44fb-6r96q\" (UID: \"a333e040-ed72-4d18-95b1-14b57a92c895\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.312004 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmj8n\" (UniqueName: \"kubernetes.io/projected/49254887-6adf-47f8-ab3b-8c8890281ee6-kube-api-access-kmj8n\") pod \"dns-default-cspmp\" (UID: \"49254887-6adf-47f8-ab3b-8c8890281ee6\") " pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.310020 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqksn\" (UniqueName: \"kubernetes.io/projected/d3064e2c-7080-4395-9a85-8d8794d588e0-kube-api-access-kqksn\") pod \"machine-approver-56656f9798-26l9f\" (UID: \"d3064e2c-7080-4395-9a85-8d8794d588e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.317590 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5c94c"] Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.329979 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/55851106-62aa-48a2-bd76-e0049671b476-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4p5wd\" (UID: \"55851106-62aa-48a2-bd76-e0049671b476\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.488082 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.491015 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.491131 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.491229 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.491789 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.491934 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.492236 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.492720 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.493333 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:49.993319701 +0000 UTC m=+181.406635748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.497478 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.497543 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.498091 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.499933 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.499965 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.500043 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.499941 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.501737 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.508907 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.519550 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 17 15:39:49 crc kubenswrapper[4767]: W0317 15:39:49.526873 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f692e30_f8c9_4823_bc82_7bf8abca0659.slice/crio-29b838ef4de1606de49092f688e71cb9ca0f2d183ed2989e390d21fb5fc977c6 WatchSource:0}: Error finding container 29b838ef4de1606de49092f688e71cb9ca0f2d183ed2989e390d21fb5fc977c6: Status 404 returned error can't find the container with id 29b838ef4de1606de49092f688e71cb9ca0f2d183ed2989e390d21fb5fc977c6 Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.528034 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.538126 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.538450 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.558965 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.568612 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.575706 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.590485 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.593412 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.593905 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:50.093880967 +0000 UTC m=+181.507197014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.597585 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.599953 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.615968 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwb2x\" (UniqueName: \"kubernetes.io/projected/4c4aa56e-7fca-4836-b8ab-581ba3be753c-kube-api-access-vwb2x\") pod \"service-ca-9c57cc56f-vwszx\" (UID: \"4c4aa56e-7fca-4836-b8ab-581ba3be753c\") " pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.628160 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.628942 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.633093 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9cz8q"] Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.639439 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.641277 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.661033 4767 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.661862 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-frl2d" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.695873 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.695939 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-webhook-cert\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.695991 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/49254887-6adf-47f8-ab3b-8c8890281ee6-metrics-tls\") pod \"dns-default-cspmp\" (UID: \"49254887-6adf-47f8-ab3b-8c8890281ee6\") " pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.696018 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-certs\") pod \"machine-config-server-bdkdk\" (UID: \"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe\") " pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.696084 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-apiservice-cert\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.696132 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49254887-6adf-47f8-ab3b-8c8890281ee6-config-volume\") pod \"dns-default-cspmp\" (UID: \"49254887-6adf-47f8-ab3b-8c8890281ee6\") " pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.696161 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509-cert\") pod \"ingress-canary-sdx96\" (UID: \"02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509\") " pod="openshift-ingress-canary/ingress-canary-sdx96" Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.696222 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:50.196207368 +0000 UTC m=+181.609523415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.696550 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-node-bootstrap-token\") pod \"machine-config-server-bdkdk\" (UID: \"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe\") " pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.698808 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.699557 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49254887-6adf-47f8-ab3b-8c8890281ee6-config-volume\") pod \"dns-default-cspmp\" (UID: \"49254887-6adf-47f8-ab3b-8c8890281ee6\") " pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.699775 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.704348 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-node-bootstrap-token\") pod \"machine-config-server-bdkdk\" (UID: \"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe\") " pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.705277 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509-cert\") pod \"ingress-canary-sdx96\" (UID: \"02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509\") " pod="openshift-ingress-canary/ingress-canary-sdx96" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.705322 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8m5cc"] Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.706268 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-apiservice-cert\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.706477 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe-certs\") pod \"machine-config-server-bdkdk\" (UID: \"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe\") " pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.710388 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b836be84-e8b0-4e8d-8bac-cfc922bea7eb-webhook-cert\") pod \"packageserver-d55dfcdfc-8rfjw\" (UID: \"b836be84-e8b0-4e8d-8bac-cfc922bea7eb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.718450 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/49254887-6adf-47f8-ab3b-8c8890281ee6-metrics-tls\") pod \"dns-default-cspmp\" (UID: \"49254887-6adf-47f8-ab3b-8c8890281ee6\") " pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.723229 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.728257 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.739147 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.746772 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.751213 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rtlwt"] Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.759200 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.764696 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.766226 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-59hsl"] Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.774512 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7smc4"] Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.797911 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:49 crc kubenswrapper[4767]: E0317 15:39:49.798633 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:50.298373515 +0000 UTC m=+181.711689572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.799351 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 17 15:39:49 crc kubenswrapper[4767]: I0317 15:39:49.806133 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.114823 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.115929 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.116151 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.116612 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:39:50 crc kubenswrapper[4767]: E0317 15:39:50.119830 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:50.619804238 +0000 UTC m=+182.033120285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.122406 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cspmp" Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.122750 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bdkdk" Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.124807 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.129610 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-sdx96" Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.196994 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw"] Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.215797 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:50 crc kubenswrapper[4767]: E0317 15:39:50.216149 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:50.716134025 +0000 UTC m=+182.129450072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.432007 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:50 crc kubenswrapper[4767]: E0317 15:39:50.432373 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:50.932360683 +0000 UTC m=+182.345676730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.468566 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" event={"ID":"8a0ddaae-cb65-45e0-a82a-44f6afb10d48","Type":"ContainerStarted","Data":"cd2135ec1df21d32437aca42407833fcaccc52595df61dc7fa36d5660bd66fa9"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.473544 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" event={"ID":"3f692e30-f8c9-4823-bc82-7bf8abca0659","Type":"ContainerStarted","Data":"29b838ef4de1606de49092f688e71cb9ca0f2d183ed2989e390d21fb5fc977c6"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.482390 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" event={"ID":"a8829d5d-100e-4518-8863-db3ab9c7b30d","Type":"ContainerStarted","Data":"d3ce379c3cde3cc810c69db45742210a401f54f019930f4f83ee0ede1038631f"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.484310 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" event={"ID":"a6576b15-805f-4a82-ba7b-1e11b5690a91","Type":"ContainerStarted","Data":"b7cc6a8ff151d85b5929a41b741dad758a9442fb26be9a7f66ee1cfc5e77a1a3"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.487602 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-59hsl" event={"ID":"471b6054-3b2a-47c9-8889-942603600da9","Type":"ContainerStarted","Data":"c79a005f6ae40613d4fb63301583078573ad8c1b001cbc4ebfb7056640e9f36d"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.489513 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ztnfb" event={"ID":"a5cf3c29-3814-4d17-876a-9ca3e44a9400","Type":"ContainerStarted","Data":"f2660bc29fc7327fc6af6c8fdfff6c759511bad1130d698b42ecac83ebe4d4d4"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.491658 4767 generic.go:334] "Generic (PLEG): container finished" podID="ece0e760-026e-4de3-bb90-bb5117963f69" containerID="ff4f2dc6547595070da2b6e3cc519d2c88b5889b8a9fe43b513ed7270ed6f899" exitCode=0 Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.491709 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" event={"ID":"ece0e760-026e-4de3-bb90-bb5117963f69","Type":"ContainerDied","Data":"ff4f2dc6547595070da2b6e3cc519d2c88b5889b8a9fe43b513ed7270ed6f899"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.504118 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" event={"ID":"2130eb2d-c288-473e-931f-68636e3afcea","Type":"ContainerStarted","Data":"2f0a92d836ef4c35482930aeb92ba06689f27ab5feeba72a31ebc1feef4f9288"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.505720 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" event={"ID":"69925f86-8108-4733-b501-e04fc8cbb320","Type":"ContainerStarted","Data":"04b3208c97c818f60bb75b67bc28c7cb9f4769e6585c9b11bcb657b197a04d4e"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.507341 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9cz8q" event={"ID":"deba5665-b026-4d87-a986-b17f02f644d7","Type":"ContainerStarted","Data":"2f730dd27ad571edc2001f46171e6a80713055b782c940d838d8f18ce435558d"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.508464 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" event={"ID":"eac52aa6-bd46-4a74-93b1-a1c299eba1c7","Type":"ContainerStarted","Data":"f457a7bb4dbd9fe48b030a5e3829af475e7918712e31685defb94a7fbd748f98"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.526738 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" event={"ID":"07a2b9f1-5291-4204-8631-c1c0f8012bc6","Type":"ContainerStarted","Data":"026f0dbb1c8b579f7f3d2734aa7b098c30015bafccc7a7d7f4f29caab3e29ece"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.528985 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" event={"ID":"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1","Type":"ContainerStarted","Data":"8aa4d12dd0d870e2f48e6ecfadf1a14cac81bc479e952be42fc44b83d06db29d"} Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.529542 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.529588 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.566031 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:50 crc kubenswrapper[4767]: E0317 15:39:50.566638 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:51.066612662 +0000 UTC m=+182.479928709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:50 crc kubenswrapper[4767]: I0317 15:39:50.947848 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:50 crc kubenswrapper[4767]: E0317 15:39:50.949346 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:51.449318695 +0000 UTC m=+182.862634812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:51 crc kubenswrapper[4767]: I0317 15:39:51.229708 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:51 crc kubenswrapper[4767]: E0317 15:39:51.230883 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:51.730854986 +0000 UTC m=+183.144171043 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:51 crc kubenswrapper[4767]: I0317 15:39:51.240379 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podStartSLOduration=117.240357419 podStartE2EDuration="1m57.240357419s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:51.127925629 +0000 UTC m=+182.541241696" watchObservedRunningTime="2026-03-17 15:39:51.240357419 +0000 UTC m=+182.653673466" Mar 17 15:39:51 crc kubenswrapper[4767]: I0317 15:39:51.336127 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:51 crc kubenswrapper[4767]: E0317 15:39:51.336524 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:51.836505342 +0000 UTC m=+183.249821389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:51 crc kubenswrapper[4767]: I0317 15:39:51.895263 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:51 crc kubenswrapper[4767]: E0317 15:39:51.895620 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:52.895604265 +0000 UTC m=+184.308920312 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:51 crc kubenswrapper[4767]: I0317 15:39:51.996379 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:51 crc kubenswrapper[4767]: E0317 15:39:51.997112 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:52.497095806 +0000 UTC m=+183.910411853 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.215328 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6"] Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.215376 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt"] Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.216382 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.216431 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:52 crc kubenswrapper[4767]: E0317 15:39:52.216898 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:52.716880509 +0000 UTC m=+184.130196566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.216921 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" event={"ID":"a90441f8-30d7-4091-a23f-2522c586f8cf","Type":"ContainerStarted","Data":"c883bbe7f75c93fb45ffd6f21675410a929bddce1f169375f941ba43c25f6251"} Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.261808 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.300090 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" event={"ID":"d3064e2c-7080-4395-9a85-8d8794d588e0","Type":"ContainerStarted","Data":"a783ad20578c9a91782f82f12c57b13beb8833a8f1859942dc963bc9c8413e93"} Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.358517 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32925427-1529-4e8d-a606-8ccbde1fd877-metrics-certs\") pod \"network-metrics-daemon-t7mfn\" (UID: \"32925427-1529-4e8d-a606-8ccbde1fd877\") " pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.361613 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:52 crc kubenswrapper[4767]: E0317 15:39:52.382814 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:52.882782291 +0000 UTC m=+184.296098338 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.382921 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:52 crc kubenswrapper[4767]: E0317 15:39:52.393088 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:52.893062275 +0000 UTC m=+184.306378322 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.502552 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:52 crc kubenswrapper[4767]: E0317 15:39:52.503862 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:53.003835654 +0000 UTC m=+184.417151701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.566972 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" event={"ID":"d333e67f-6e87-4ce6-8d3f-4ff16caf2159","Type":"ContainerStarted","Data":"d009523c19a424595ca6056acc2a3789a1be3e1ea755462e646b89cac1b02204"} Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.569927 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.577648 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t7mfn" Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.626813 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:52 crc kubenswrapper[4767]: E0317 15:39:52.638866 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:53.138848414 +0000 UTC m=+184.552164461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.738979 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:52 crc kubenswrapper[4767]: E0317 15:39:52.744213 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:53.244165353 +0000 UTC m=+184.657481400 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.808868 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6"] Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.815801 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5"] Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.841348 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:52 crc kubenswrapper[4767]: E0317 15:39:52.841878 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:53.341859607 +0000 UTC m=+184.755175654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.944720 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.951266 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w"] Mar 17 15:39:52 crc kubenswrapper[4767]: I0317 15:39:52.960346 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq"] Mar 17 15:39:52 crc kubenswrapper[4767]: W0317 15:39:52.980833 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a6bb9ed_1480_4ecf_a982_d36eafb4ddfe.slice/crio-99341509056e68a941b7acc8ba14571187bfe94888e5bdab83454822e9ae80f1 WatchSource:0}: Error finding container 99341509056e68a941b7acc8ba14571187bfe94888e5bdab83454822e9ae80f1: Status 404 returned error can't find the container with id 99341509056e68a941b7acc8ba14571187bfe94888e5bdab83454822e9ae80f1 Mar 17 15:39:52 crc kubenswrapper[4767]: E0317 15:39:52.994576 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:53.49453473 +0000 UTC m=+184.907850777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:53 crc kubenswrapper[4767]: I0317 15:39:53.032214 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7dzwj" podStartSLOduration=118.032186535 podStartE2EDuration="1m58.032186535s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:53.018995567 +0000 UTC m=+184.432311624" watchObservedRunningTime="2026-03-17 15:39:53.032186535 +0000 UTC m=+184.445502582" Mar 17 15:39:53 crc kubenswrapper[4767]: I0317 15:39:53.046941 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:53 crc kubenswrapper[4767]: E0317 15:39:53.047337 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:53.547324543 +0000 UTC m=+184.960640590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:53 crc kubenswrapper[4767]: I0317 15:39:53.150645 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:53 crc kubenswrapper[4767]: E0317 15:39:53.151107 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:53.651091823 +0000 UTC m=+185.064407870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:53 crc kubenswrapper[4767]: I0317 15:39:53.259764 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:53 crc kubenswrapper[4767]: E0317 15:39:53.260199 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:53.760183549 +0000 UTC m=+185.173499596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:53 crc kubenswrapper[4767]: I0317 15:39:53.444148 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:53 crc kubenswrapper[4767]: E0317 15:39:53.444633 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:53.944603315 +0000 UTC m=+185.357919362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:53 crc kubenswrapper[4767]: I0317 15:39:53.552001 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:53 crc kubenswrapper[4767]: E0317 15:39:53.552841 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:54.052658914 +0000 UTC m=+185.465974961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:53 crc kubenswrapper[4767]: I0317 15:39:53.698028 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:53 crc kubenswrapper[4767]: E0317 15:39:53.698404 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:54.198386967 +0000 UTC m=+185.611703014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:53 crc kubenswrapper[4767]: I0317 15:39:53.799832 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:53 crc kubenswrapper[4767]: E0317 15:39:53.800131 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:54.300119942 +0000 UTC m=+185.713435989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:53.998608 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:54 crc kubenswrapper[4767]: E0317 15:39:53.999140 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:54.499123959 +0000 UTC m=+185.912440006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.000043 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" event={"ID":"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1","Type":"ContainerStarted","Data":"c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.000527 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.008231 4767 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8m5cc container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.008271 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" podUID="0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.040506 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" event={"ID":"eac52aa6-bd46-4a74-93b1-a1c299eba1c7","Type":"ContainerStarted","Data":"8e943dfb4e3282068aebf8bf2cdb5363edb34b08a5706d56557612a2ea4b0531"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.050099 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ztnfb" event={"ID":"a5cf3c29-3814-4d17-876a-9ca3e44a9400","Type":"ContainerStarted","Data":"3dbc65c86754010059515f8fc9873782ca67f8577939e37319a0dcd29bf71113"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.052493 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" event={"ID":"68f6ff1d-a119-4faf-89aa-fd9afc2d93f8","Type":"ContainerStarted","Data":"5989feb205599a5d1783ab649454e7cca88b48a32c815ad19f0e79dd8c1988be"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.055531 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" event={"ID":"3f692e30-f8c9-4823-bc82-7bf8abca0659","Type":"ContainerStarted","Data":"dad7ccdb9a65184f778ab73bb3462847e4dd429848a1f0f8af7f918cb1ffeef1"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.066047 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9cz8q" event={"ID":"deba5665-b026-4d87-a986-b17f02f644d7","Type":"ContainerStarted","Data":"415bc454f7b25622d72c606b86b13b8ba7e778cd5cef1a1a64a9fb92f886fbf6"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.068760 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" event={"ID":"69925f86-8108-4733-b501-e04fc8cbb320","Type":"ContainerStarted","Data":"17e3ca0cd39523398f589a3519de6665965f686257886291ce91f4c606eac12d"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.078389 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" event={"ID":"a8829d5d-100e-4518-8863-db3ab9c7b30d","Type":"ContainerStarted","Data":"3942c07c07c5367fe53baeedc2dbb54cea88cdea4ebf87f7f586fc0bf82810ae"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.080847 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bdkdk" event={"ID":"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe","Type":"ContainerStarted","Data":"99341509056e68a941b7acc8ba14571187bfe94888e5bdab83454822e9ae80f1"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.082451 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" event={"ID":"2dcf5766-ea05-4364-bd69-c51987442223","Type":"ContainerStarted","Data":"5dab89be1c352d5079e53bacee5d520f8533c0d4fc74e81c137209c0d74cc786"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.083930 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" event={"ID":"a6576b15-805f-4a82-ba7b-1e11b5690a91","Type":"ContainerStarted","Data":"48fe9cfd5e579e3706075d286aa251a5d134b5d8f5831a8b59ad8cd20df5b1de"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.095982 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-59hsl" event={"ID":"471b6054-3b2a-47c9-8889-942603600da9","Type":"ContainerStarted","Data":"c82a98015caa0bb7c218c4c3adb06778f84051c358433937ea7560b971a25552"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.099113 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-59hsl" Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.107452 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" event={"ID":"b896af46-1936-4488-9def-e860440a98bf","Type":"ContainerStarted","Data":"c1908379b7e6de2f7e6eae18592e3bbf1ea58c838fef39af04c3ef2045bdcce4"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.109630 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.109692 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.132300 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" event={"ID":"429dd8b3-15c5-4ed5-a81f-1c0f14c1434c","Type":"ContainerStarted","Data":"1b5542b0d60f0aae504785fbd34ee41863ca4dbaf7db1938646db9995d82b3b7"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.154811 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:54 crc kubenswrapper[4767]: E0317 15:39:54.159930 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:54.659914918 +0000 UTC m=+186.073230965 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.162138 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" event={"ID":"07a2b9f1-5291-4204-8631-c1c0f8012bc6","Type":"ContainerStarted","Data":"9190aea82600d49a869bd59e55d252c396447b1b2ac38153985bcd2192517f1e"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.167233 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" event={"ID":"9b910672-1706-4ce5-8871-e006b25c513d","Type":"ContainerStarted","Data":"e783e3c41b637c9b23c234858a9335bb22309a2b038afdf58c62012d134180a5"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.187635 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w" event={"ID":"246c8a34-7be4-4e94-ab60-c4e471a762b1","Type":"ContainerStarted","Data":"1b932b90ffa05bdec3986b33f0790ddd719d9bd5c707a838a4b005c429a722f2"} Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.291074 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.291971 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:54 crc kubenswrapper[4767]: E0317 15:39:54.292274 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:54.792243457 +0000 UTC m=+186.205559504 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.295153 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:54 crc kubenswrapper[4767]: E0317 15:39:54.295629 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:54.795613584 +0000 UTC m=+186.208929631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.297529 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" podStartSLOduration=120.297499022 podStartE2EDuration="2m0.297499022s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:54.155043033 +0000 UTC m=+185.568359090" watchObservedRunningTime="2026-03-17 15:39:54.297499022 +0000 UTC m=+185.710815079" Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.302735 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-9cz8q" podStartSLOduration=120.302714116 podStartE2EDuration="2m0.302714116s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:54.291592821 +0000 UTC m=+185.704908878" watchObservedRunningTime="2026-03-17 15:39:54.302714116 +0000 UTC m=+185.716030173" Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.321422 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bk7p5"] Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.324185 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms"] Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.327700 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c4mkn" podStartSLOduration=120.327644474 podStartE2EDuration="2m0.327644474s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:54.317444063 +0000 UTC m=+185.730760110" watchObservedRunningTime="2026-03-17 15:39:54.327644474 +0000 UTC m=+185.740960541" Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.412589 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.413470 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-59hsl" podStartSLOduration=120.413454662 podStartE2EDuration="2m0.413454662s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:54.412278682 +0000 UTC m=+185.825594729" watchObservedRunningTime="2026-03-17 15:39:54.413454662 +0000 UTC m=+185.826770709" Mar 17 15:39:54 crc kubenswrapper[4767]: E0317 15:39:54.413643 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:54.913612006 +0000 UTC m=+186.326928063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.475716 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" podStartSLOduration=120.475690646 podStartE2EDuration="2m0.475690646s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:54.473972702 +0000 UTC m=+185.887288769" watchObservedRunningTime="2026-03-17 15:39:54.475690646 +0000 UTC m=+185.889006693" Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.515844 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:54 crc kubenswrapper[4767]: E0317 15:39:54.524551 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:55.024532297 +0000 UTC m=+186.437848344 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.700855 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:54 crc kubenswrapper[4767]: E0317 15:39:54.708671 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:55.208634293 +0000 UTC m=+186.621950330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.741841 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-ztnfb" podStartSLOduration=120.741818913 podStartE2EDuration="2m0.741818913s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:54.729827355 +0000 UTC m=+186.143143422" watchObservedRunningTime="2026-03-17 15:39:54.741818913 +0000 UTC m=+186.155134970" Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.802847 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.803217 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.808279 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:54 crc kubenswrapper[4767]: E0317 15:39:54.809201 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:55.309155297 +0000 UTC m=+186.722471364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:54 crc kubenswrapper[4767]: I0317 15:39:54.933230 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:54 crc kubenswrapper[4767]: E0317 15:39:54.933700 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:55.433685097 +0000 UTC m=+186.847001144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.039784 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:55 crc kubenswrapper[4767]: E0317 15:39:55.040147 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:55.540132203 +0000 UTC m=+186.953448250 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.062741 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcqcq" podStartSLOduration=121.062719372 podStartE2EDuration="2m1.062719372s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:54.943803266 +0000 UTC m=+186.357119313" watchObservedRunningTime="2026-03-17 15:39:55.062719372 +0000 UTC m=+186.476035419" Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.146049 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:55 crc kubenswrapper[4767]: E0317 15:39:55.146430 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:55.646414846 +0000 UTC m=+187.059730893 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.247347 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:55 crc kubenswrapper[4767]: E0317 15:39:55.248105 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:55.748089129 +0000 UTC m=+187.161405176 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.278686 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.278762 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.373236 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:55 crc kubenswrapper[4767]: E0317 15:39:55.374600 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:55.874576939 +0000 UTC m=+187.287892986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.396409 4767 generic.go:334] "Generic (PLEG): container finished" podID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerID="dad7ccdb9a65184f778ab73bb3462847e4dd429848a1f0f8af7f918cb1ffeef1" exitCode=0 Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.396763 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" event={"ID":"3f692e30-f8c9-4823-bc82-7bf8abca0659","Type":"ContainerDied","Data":"dad7ccdb9a65184f778ab73bb3462847e4dd429848a1f0f8af7f918cb1ffeef1"} Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.401651 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" event={"ID":"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3","Type":"ContainerStarted","Data":"20013c43198a556510496ff2bc8ee5f368120c437023736b7475efc3e20d9b40"} Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.419693 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" event={"ID":"1c88a6ee-cd24-4d85-8a89-d830e5baa434","Type":"ContainerStarted","Data":"13c924bcdd6ec6c7ace5d14f78e3f1903f38f616cf3b0281ca8c28bf9279cf6f"} Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.422727 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.422764 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.422884 4767 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8m5cc container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.422982 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" podUID="0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.481233 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:55 crc kubenswrapper[4767]: E0317 15:39:55.482933 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:55.982919814 +0000 UTC m=+187.396235861 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.525204 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-5c94c" podStartSLOduration=121.525187336 podStartE2EDuration="2m1.525187336s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:55.243091751 +0000 UTC m=+186.656407808" watchObservedRunningTime="2026-03-17 15:39:55.525187336 +0000 UTC m=+186.938503373" Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.527248 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-cwz8x"] Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.576167 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4"] Mar 17 15:39:55 crc kubenswrapper[4767]: W0317 15:39:55.579717 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2047989e_3a86_4a4e_8b4a_c8d29fc17e34.slice/crio-e58d5c3a0d7e94cf20fd7cf43fbafe6daaa07e1eb07740e06482d0d9c6b8e115 WatchSource:0}: Error finding container e58d5c3a0d7e94cf20fd7cf43fbafe6daaa07e1eb07740e06482d0d9c6b8e115: Status 404 returned error can't find the container with id e58d5c3a0d7e94cf20fd7cf43fbafe6daaa07e1eb07740e06482d0d9c6b8e115 Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.584845 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:55 crc kubenswrapper[4767]: E0317 15:39:55.585285 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:56.085265905 +0000 UTC m=+187.498581962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.608720 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q"] Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.629443 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd"] Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.659181 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f"] Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.686792 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:55 crc kubenswrapper[4767]: E0317 15:39:55.687361 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:56.1873431 +0000 UTC m=+187.600659147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.697902 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt"] Mar 17 15:39:55 crc kubenswrapper[4767]: I0317 15:39:55.708863 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cb6sk"] Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:55.842059 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:56 crc kubenswrapper[4767]: E0317 15:39:55.842928 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:56.342901684 +0000 UTC m=+187.756217731 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:55.943574 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:56 crc kubenswrapper[4767]: E0317 15:39:55.944095 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:56.444079075 +0000 UTC m=+187.857395122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.044112 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:56 crc kubenswrapper[4767]: E0317 15:39:56.044495 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:56.544473687 +0000 UTC m=+187.957789734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.148221 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:56 crc kubenswrapper[4767]: E0317 15:39:56.148624 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:56.648608004 +0000 UTC m=+188.061924051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.249137 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:56 crc kubenswrapper[4767]: E0317 15:39:56.249519 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:56.749504319 +0000 UTC m=+188.162820366 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.250289 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cspmp"] Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.256960 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vwszx"] Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.259613 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr"] Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.264821 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz"] Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.266276 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw"] Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.271803 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx"] Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.273504 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:39:56 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:39:56 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:39:56 crc kubenswrapper[4767]: healthz check failed Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.273552 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.274332 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-frl2d"] Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.278600 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-sdx96"] Mar 17 15:39:56 crc kubenswrapper[4767]: W0317 15:39:56.291969 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49254887_6adf_47f8_ab3b_8c8890281ee6.slice/crio-933962c5c906238cfab87d2e0d8cc2d72848f410ff9e0d07a336d15277336fb0 WatchSource:0}: Error finding container 933962c5c906238cfab87d2e0d8cc2d72848f410ff9e0d07a336d15277336fb0: Status 404 returned error can't find the container with id 933962c5c906238cfab87d2e0d8cc2d72848f410ff9e0d07a336d15277336fb0 Mar 17 15:39:56 crc kubenswrapper[4767]: W0317 15:39:56.298652 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c4aa56e_7fca_4836_b8ab_581ba3be753c.slice/crio-fb1583bf9661ce09e08225680f5855f135536f1ef06781cd2eb6e58eb560d592 WatchSource:0}: Error finding container fb1583bf9661ce09e08225680f5855f135536f1ef06781cd2eb6e58eb560d592: Status 404 returned error can't find the container with id fb1583bf9661ce09e08225680f5855f135536f1ef06781cd2eb6e58eb560d592 Mar 17 15:39:56 crc kubenswrapper[4767]: W0317 15:39:56.314478 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26899a2e_90e1_49e1_a86b_bfa19c62c195.slice/crio-c51ef7153464b8214d28965e6ba192683340b10784efc21601c5a0fdb5c4f656 WatchSource:0}: Error finding container c51ef7153464b8214d28965e6ba192683340b10784efc21601c5a0fdb5c4f656: Status 404 returned error can't find the container with id c51ef7153464b8214d28965e6ba192683340b10784efc21601c5a0fdb5c4f656 Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.352033 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:56 crc kubenswrapper[4767]: E0317 15:39:56.353098 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:56.853075082 +0000 UTC m=+188.266391299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.360049 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-t7mfn"] Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.451030 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" event={"ID":"4c4aa56e-7fca-4836-b8ab-581ba3be753c","Type":"ContainerStarted","Data":"fb1583bf9661ce09e08225680f5855f135536f1ef06781cd2eb6e58eb560d592"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.453231 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" event={"ID":"55851106-62aa-48a2-bd76-e0049671b476","Type":"ContainerStarted","Data":"e2987889e83800808dec0cf8e0322c786d08ded8a83c4739952d21aa95d913bd"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.453957 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:56 crc kubenswrapper[4767]: E0317 15:39:56.454340 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:56.954320835 +0000 UTC m=+188.367636882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.465827 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bdkdk" event={"ID":"3a6bb9ed-1480-4ecf-a982-d36eafb4ddfe","Type":"ContainerStarted","Data":"e9b410c12eb9d8d1db57a4f42bbecc0dc607041b535e3675c9f625bbcdeff689"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.477360 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" event={"ID":"b896af46-1936-4488-9def-e860440a98bf","Type":"ContainerStarted","Data":"5ef509fe2fbd1e28c5fa648e3b2ccd296698d3b5724088421e4ecc517f98b9aa"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.480473 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" event={"ID":"a90441f8-30d7-4091-a23f-2522c586f8cf","Type":"ContainerStarted","Data":"3efe3ea49c08258f3ba548a8ea7c07dce8d7dd6c72099fdb81abb6d455a891a9"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.481469 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.487946 4767 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-tz4lw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.488493 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" podUID="a90441f8-30d7-4091-a23f-2522c586f8cf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.489400 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" event={"ID":"68f6ff1d-a119-4faf-89aa-fd9afc2d93f8","Type":"ContainerStarted","Data":"2aebe6ce8cabdfcaed25d3542ad12e9da57821b51fef0210e0be3b8fd010eca5"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.531478 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" event={"ID":"bbbc3925-9201-47fa-a29f-e6b271de61fd","Type":"ContainerStarted","Data":"419120ffa0802ace28a6b18b6c5e3cf8d474bb66bd317e09de3441658a595a4a"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.533347 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz" event={"ID":"26899a2e-90e1-49e1-a86b-bfa19c62c195","Type":"ContainerStarted","Data":"c51ef7153464b8214d28965e6ba192683340b10784efc21601c5a0fdb5c4f656"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.534250 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-frl2d" event={"ID":"201932aa-c539-434b-992e-92f889e52de3","Type":"ContainerStarted","Data":"92dd326e52f9de4e54e2fba43c5115a4bfacedd195fa9da3f868504cac2bd736"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.535110 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" event={"ID":"32925427-1529-4e8d-a606-8ccbde1fd877","Type":"ContainerStarted","Data":"92cfa4ba4354f39d711de175a6682102044b89970eefadfe37c7fdce3da48519"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.535899 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" event={"ID":"acdeabaa-85fb-4d85-bad6-89fc1be0125a","Type":"ContainerStarted","Data":"13fb5834ff825d4c4df1f6eff8c22425de2e947991d4fc85ab45dbfae8310bec"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.536893 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" event={"ID":"8a0ddaae-cb65-45e0-a82a-44f6afb10d48","Type":"ContainerStarted","Data":"a9715511abf3e82c7c36b395096eec861e4c5f36b40fa5921a952218d7781141"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.537866 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.539018 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" event={"ID":"429dd8b3-15c5-4ed5-a81f-1c0f14c1434c","Type":"ContainerStarted","Data":"377f4348f44bc71318e73586740e2f436dca6a20db51e35464e0467a6e3e3eac"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.540804 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" event={"ID":"b836be84-e8b0-4e8d-8bac-cfc922bea7eb","Type":"ContainerStarted","Data":"f01bc64c22fc270b4803dcaacf5e0dc9bdd33b3a4160e2afb7aff3cd6ceae435"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.542996 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" event={"ID":"a333e040-ed72-4d18-95b1-14b57a92c895","Type":"ContainerStarted","Data":"c7e78dbec517bf9fd49edbb4d265ed27ceb072a3cb84fb5cb15b585280e28f25"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.543961 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" event={"ID":"2047989e-3a86-4a4e-8b4a-c8d29fc17e34","Type":"ContainerStarted","Data":"e58d5c3a0d7e94cf20fd7cf43fbafe6daaa07e1eb07740e06482d0d9c6b8e115"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.546385 4767 generic.go:334] "Generic (PLEG): container finished" podID="1c88a6ee-cd24-4d85-8a89-d830e5baa434" containerID="723e5b26aa23d74f0bfe990597041fa559d91ef31811b357d6f72aa4bcea9aec" exitCode=0 Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.546433 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" event={"ID":"1c88a6ee-cd24-4d85-8a89-d830e5baa434","Type":"ContainerDied","Data":"723e5b26aa23d74f0bfe990597041fa559d91ef31811b357d6f72aa4bcea9aec"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.550842 4767 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rtlwt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.550959 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" podUID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.551200 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" event={"ID":"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc","Type":"ContainerStarted","Data":"faed4d92fa5b4a362d0f03b11d99bc741fc8dd24fc219ac197f3b20554f63dbd"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.555067 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:56 crc kubenswrapper[4767]: E0317 15:39:56.556292 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:57.056280036 +0000 UTC m=+188.469596083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.635149 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" event={"ID":"d3064e2c-7080-4395-9a85-8d8794d588e0","Type":"ContainerStarted","Data":"fccd80d230adf5801e686de0c3d2a71fb91c7eb88115663e7d2710e2887a425d"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.763726 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-sdx96" event={"ID":"02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509","Type":"ContainerStarted","Data":"6d32799ffd30fa7f85f21889d3e731fe4209ab738c64a6b192cd9c126fcd228e"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.764885 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bdkdk" podStartSLOduration=10.764869009 podStartE2EDuration="10.764869009s" podCreationTimestamp="2026-03-17 15:39:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:56.764273954 +0000 UTC m=+188.177590011" watchObservedRunningTime="2026-03-17 15:39:56.764869009 +0000 UTC m=+188.178185076" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.765615 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:56 crc kubenswrapper[4767]: E0317 15:39:56.765935 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:57.265911116 +0000 UTC m=+188.679227163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.776031 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" event={"ID":"d333e67f-6e87-4ce6-8d3f-4ff16caf2159","Type":"ContainerStarted","Data":"8671318adead2d0b1caded0ccdfd667b1dfb71c09dee51d4b2cd82f2bd35fc32"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.793703 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" event={"ID":"3398275a-5195-45d6-9444-9e4d9825770f","Type":"ContainerStarted","Data":"916a93d8dcc8e416149485d7ff9e1c9d4b9888798ebeb307754c074d3d9ba479"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.806799 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cspmp" event={"ID":"49254887-6adf-47f8-ab3b-8c8890281ee6","Type":"ContainerStarted","Data":"933962c5c906238cfab87d2e0d8cc2d72848f410ff9e0d07a336d15277336fb0"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.810768 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" event={"ID":"75f035cd-32d9-44a2-8d8f-3d5f40742b67","Type":"ContainerStarted","Data":"3ddaa1aa6d138ce0e49f52cf1a5decfc2908d8046f4221b415b8ebcee6389eff"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.811980 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" event={"ID":"9b910672-1706-4ce5-8871-e006b25c513d","Type":"ContainerStarted","Data":"7a949f711481e2daaf660777c592d3bd0d65cef343d2df87a60d6e3d82c0d9fb"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.814723 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" event={"ID":"3f692e30-f8c9-4823-bc82-7bf8abca0659","Type":"ContainerStarted","Data":"0f8553bfc41dec31090eb663209a1e4657345284d2f6241ad40d2b406f04d846"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.815514 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.871946 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:56 crc kubenswrapper[4767]: E0317 15:39:56.874746 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:57.374720263 +0000 UTC m=+188.788036310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.882679 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" event={"ID":"ece0e760-026e-4de3-bb90-bb5117963f69","Type":"ContainerStarted","Data":"0c20e9e43d15d8cb76092fcfca87eee778579238cacc303fc4a545632a0d0502"} Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.893360 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.893431 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.932199 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.963200 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.963274 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.978459 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:56 crc kubenswrapper[4767]: E0317 15:39:56.978553 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:57.478537332 +0000 UTC m=+188.891853379 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.978967 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.980737 4767 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-pv6ds container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 17 15:39:56 crc kubenswrapper[4767]: I0317 15:39:56.980782 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" podUID="ece0e760-026e-4de3-bb90-bb5117963f69" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 17 15:39:57 crc kubenswrapper[4767]: E0317 15:39:57.103804 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:57.60378188 +0000 UTC m=+189.017097927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.312129 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:57 crc kubenswrapper[4767]: E0317 15:39:57.315889 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:57.815864052 +0000 UTC m=+189.229180099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.321312 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qdjwq" podStartSLOduration=122.321292511 podStartE2EDuration="2m2.321292511s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:57.310028542 +0000 UTC m=+188.723344599" watchObservedRunningTime="2026-03-17 15:39:57.321292511 +0000 UTC m=+188.734608558" Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.499596 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:39:57 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:39:57 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:39:57 crc kubenswrapper[4767]: healthz check failed Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.505259 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.538159 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:57 crc kubenswrapper[4767]: E0317 15:39:57.540353 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:58.040335911 +0000 UTC m=+189.453651968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.639781 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:57 crc kubenswrapper[4767]: E0317 15:39:57.640248 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:58.140233 +0000 UTC m=+189.553549047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.743401 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:57 crc kubenswrapper[4767]: E0317 15:39:57.743961 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:58.243947717 +0000 UTC m=+189.657263764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.781500 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2kkt6" podStartSLOduration=122.781480158 podStartE2EDuration="2m2.781480158s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:57.775874624 +0000 UTC m=+189.189190671" watchObservedRunningTime="2026-03-17 15:39:57.781480158 +0000 UTC m=+189.194796205" Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.834266 4767 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rtlwt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.834330 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" podUID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.835580 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" podStartSLOduration=123.835571493 podStartE2EDuration="2m3.835571493s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:57.835308226 +0000 UTC m=+189.248624283" watchObservedRunningTime="2026-03-17 15:39:57.835571493 +0000 UTC m=+189.248887540" Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.844835 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:57 crc kubenswrapper[4767]: E0317 15:39:57.846863 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:58.346826952 +0000 UTC m=+189.760142999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.853217 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:57 crc kubenswrapper[4767]: E0317 15:39:57.853962 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:58.353944194 +0000 UTC m=+189.767260241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.864746 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" podStartSLOduration=122.86472747 podStartE2EDuration="2m2.86472747s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:57.861312543 +0000 UTC m=+189.274628610" watchObservedRunningTime="2026-03-17 15:39:57.86472747 +0000 UTC m=+189.278043517" Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.881949 4767 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8m5cc container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.882045 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" podUID="0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.937234 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" event={"ID":"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3","Type":"ContainerStarted","Data":"5ae467dd1d964c180822005639e5a6b467d4b56f01e3b70a9bf2396e3a90c3d1"} Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.937956 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" podStartSLOduration=122.937931405 podStartE2EDuration="2m2.937931405s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:57.936369005 +0000 UTC m=+189.349685052" watchObservedRunningTime="2026-03-17 15:39:57.937931405 +0000 UTC m=+189.351247472" Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.945131 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" event={"ID":"69925f86-8108-4733-b501-e04fc8cbb320","Type":"ContainerStarted","Data":"7eb2e7c61177c1d205dafdebfa13671a72bfd581c61fe4d94afa54c9c0ad480d"} Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.955496 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:57 crc kubenswrapper[4767]: E0317 15:39:57.956765 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:58.456743547 +0000 UTC m=+189.870059624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.961928 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" event={"ID":"2dcf5766-ea05-4364-bd69-c51987442223","Type":"ContainerStarted","Data":"4ee03f44c845164e4b88b3a9fa82dc161170e0e03bf3da10b72e49f39b65494e"} Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.981372 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w" event={"ID":"246c8a34-7be4-4e94-ab60-c4e471a762b1","Type":"ContainerStarted","Data":"fe9b256da3617d73b6140f653026ba4e20f92d141fdbbd969ae36889a0261669"} Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.984245 4767 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-tz4lw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.984293 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" podUID="a90441f8-30d7-4091-a23f-2522c586f8cf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.984939 4767 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rtlwt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.984963 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" podUID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 17 15:39:57 crc kubenswrapper[4767]: I0317 15:39:57.992869 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dq9tt" podStartSLOduration=122.992851222 podStartE2EDuration="2m2.992851222s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:57.990326817 +0000 UTC m=+189.403642874" watchObservedRunningTime="2026-03-17 15:39:57.992851222 +0000 UTC m=+189.406167269" Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.018567 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.019858 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.020014 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podStartSLOduration=124.020004557 podStartE2EDuration="2m4.020004557s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:58.019037352 +0000 UTC m=+189.432353419" watchObservedRunningTime="2026-03-17 15:39:58.020004557 +0000 UTC m=+189.433320604" Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.042254 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plft5" podStartSLOduration=123.042238727 podStartE2EDuration="2m3.042238727s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:58.041229511 +0000 UTC m=+189.454545568" watchObservedRunningTime="2026-03-17 15:39:58.042238727 +0000 UTC m=+189.455554774" Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.046332 4767 patch_prober.go:28] interesting pod/console-f9d7485db-9cz8q container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.046374 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9cz8q" podUID="deba5665-b026-4d87-a986-b17f02f644d7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.057348 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:58 crc kubenswrapper[4767]: E0317 15:39:58.059000 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:58.558979936 +0000 UTC m=+189.972295983 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.159876 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:58 crc kubenswrapper[4767]: E0317 15:39:58.160924 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:58.660899036 +0000 UTC m=+190.074215083 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.270815 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.271094 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.271144 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:39:58 crc kubenswrapper[4767]: E0317 15:39:58.271418 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:58.771400586 +0000 UTC m=+190.184716633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.272398 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.272433 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.282329 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:39:58 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:39:58 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:39:58 crc kubenswrapper[4767]: healthz check failed Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.282414 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.283132 4767 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-tz4lw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.283165 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" podUID="a90441f8-30d7-4091-a23f-2522c586f8cf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.380728 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:58 crc kubenswrapper[4767]: E0317 15:39:58.381054 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:58.881020504 +0000 UTC m=+190.294336551 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.381277 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:58 crc kubenswrapper[4767]: E0317 15:39:58.381710 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:58.881693221 +0000 UTC m=+190.295009258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.548465 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:58 crc kubenswrapper[4767]: E0317 15:39:58.549250 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:59.049157951 +0000 UTC m=+190.462474008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.677893 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:58 crc kubenswrapper[4767]: E0317 15:39:58.678447 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:59.178427371 +0000 UTC m=+190.591743418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.785901 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:58 crc kubenswrapper[4767]: E0317 15:39:58.786327 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:59.286312495 +0000 UTC m=+190.699628532 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:58 crc kubenswrapper[4767]: I0317 15:39:58.992410 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:58 crc kubenswrapper[4767]: E0317 15:39:58.993201 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:59.493188563 +0000 UTC m=+190.906504610 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.115112 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:59 crc kubenswrapper[4767]: E0317 15:39:59.115389 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:59.615373382 +0000 UTC m=+191.028689429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.160409 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.160476 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.160495 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.160528 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.245432 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" event={"ID":"d333e67f-6e87-4ce6-8d3f-4ff16caf2159","Type":"ContainerStarted","Data":"1d96e7e8e41757940162834ef18b5ec36c5ea7b897ff449f7e9cd659ed4052c6"} Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.361856 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.366464 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:59 crc kubenswrapper[4767]: E0317 15:39:59.367023 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:39:59.867005008 +0000 UTC m=+191.280321055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.367189 4767 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8m5cc container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.367231 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" podUID="0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.369026 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:39:59 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:39:59 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:39:59 crc kubenswrapper[4767]: healthz check failed Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.369104 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.417384 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-lfctz" podStartSLOduration=125.417369938 podStartE2EDuration="2m5.417369938s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:58.066433656 +0000 UTC m=+189.479749703" watchObservedRunningTime="2026-03-17 15:39:59.417369938 +0000 UTC m=+190.830685985" Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.470269 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:59 crc kubenswrapper[4767]: E0317 15:39:59.471900 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:39:59.971875784 +0000 UTC m=+191.385191831 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.548962 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" event={"ID":"55851106-62aa-48a2-bd76-e0049671b476","Type":"ContainerStarted","Data":"c6f79a80a876739b907504c0fd0f9c2887db49820f7cfff3c805fa913f64069c"} Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.576078 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:39:59 crc kubenswrapper[4767]: E0317 15:39:59.576573 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:00.076560435 +0000 UTC m=+191.489876482 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.746989 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:39:59 crc kubenswrapper[4767]: E0317 15:39:59.748445 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:00.248420467 +0000 UTC m=+191.661736514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.748681 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" event={"ID":"acdeabaa-85fb-4d85-bad6-89fc1be0125a","Type":"ContainerStarted","Data":"eae33d5c864c011827785ea5feaf0c9e88f81a1c3da2b35d6871e80fab77a840"} Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.749562 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.749602 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.749673 4767 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-tz4lw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.749687 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" podUID="a90441f8-30d7-4091-a23f-2522c586f8cf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.750099 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.750178 4767 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rtlwt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.750208 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" podUID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.810064 4767 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cb6sk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 15:39:59 crc kubenswrapper[4767]: I0317 15:39:59.810126 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" podUID="acdeabaa-85fb-4d85-bad6-89fc1be0125a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:39:59.848184 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:00 crc kubenswrapper[4767]: E0317 15:39:59.854610 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:00.354591037 +0000 UTC m=+191.767907164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:39:59.888594 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-7smc4" podStartSLOduration=124.888574597 podStartE2EDuration="2m4.888574597s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:39:59.418372933 +0000 UTC m=+190.831688980" watchObservedRunningTime="2026-03-17 15:39:59.888574597 +0000 UTC m=+191.301890644" Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:39:59.977044 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:00 crc kubenswrapper[4767]: E0317 15:39:59.977587 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:00.477565936 +0000 UTC m=+191.890881983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.264644 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:00 crc kubenswrapper[4767]: E0317 15:40:00.265068 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:00.7650485 +0000 UTC m=+192.178364547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.276624 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:00 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:00 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:00 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.276673 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.494963 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:00 crc kubenswrapper[4767]: E0317 15:40:00.495383 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:00.995368109 +0000 UTC m=+192.408684156 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.683055 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562700-h4fl7"] Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.684731 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562700-h4fl7" Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.769897 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.769971 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.770947 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.771012 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmjcq\" (UniqueName: \"kubernetes.io/projected/051f576f-ce90-4401-8944-a340dd67b274-kube-api-access-wmjcq\") pod \"auto-csr-approver-29562700-h4fl7\" (UID: \"051f576f-ce90-4401-8944-a340dd67b274\") " pod="openshift-infra/auto-csr-approver-29562700-h4fl7" Mar 17 15:40:00 crc kubenswrapper[4767]: E0317 15:40:00.771463 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:01.271450491 +0000 UTC m=+192.684766538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.880049 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" event={"ID":"32925427-1529-4e8d-a606-8ccbde1fd877","Type":"ContainerStarted","Data":"f2aeabb0bd39d0a43620a4face3fffe664fba99b85ad9f54378974c44402d7d7"} Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.881009 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.881200 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmjcq\" (UniqueName: \"kubernetes.io/projected/051f576f-ce90-4401-8944-a340dd67b274-kube-api-access-wmjcq\") pod \"auto-csr-approver-29562700-h4fl7\" (UID: \"051f576f-ce90-4401-8944-a340dd67b274\") " pod="openshift-infra/auto-csr-approver-29562700-h4fl7" Mar 17 15:40:00 crc kubenswrapper[4767]: E0317 15:40:00.881658 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:01.381641353 +0000 UTC m=+192.794957400 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.884731 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" event={"ID":"a333e040-ed72-4d18-95b1-14b57a92c895","Type":"ContainerStarted","Data":"52a6acfe67114f9acc7cfa43c05c4e8be607930c51ff296233e402842a460fb6"} Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.885749 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.887796 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" event={"ID":"75f035cd-32d9-44a2-8d8f-3d5f40742b67","Type":"ContainerStarted","Data":"59c17ad5134f8551f9dd65e38fcb77e2bca464ade2812471744070029473321e"} Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.888673 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:40:00 crc kubenswrapper[4767]: I0317 15:40:00.924629 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562700-h4fl7"] Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.040109 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.042543 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.042604 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 17 15:40:01 crc kubenswrapper[4767]: E0317 15:40:01.042651 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:01.542632247 +0000 UTC m=+192.955948294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.042695 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.042711 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.050308 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz" event={"ID":"26899a2e-90e1-49e1-a86b-bfa19c62c195","Type":"ContainerStarted","Data":"a59a4a303ba9a48da3f73e082984200047f20fddb03292b56583233bf9e821c4"} Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.187691 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmjcq\" (UniqueName: \"kubernetes.io/projected/051f576f-ce90-4401-8944-a340dd67b274-kube-api-access-wmjcq\") pod \"auto-csr-approver-29562700-h4fl7\" (UID: \"051f576f-ce90-4401-8944-a340dd67b274\") " pod="openshift-infra/auto-csr-approver-29562700-h4fl7" Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.196058 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:01 crc kubenswrapper[4767]: E0317 15:40:01.196536 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:01.696516898 +0000 UTC m=+193.109832955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.196685 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:01 crc kubenswrapper[4767]: E0317 15:40:01.198213 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:01.698196111 +0000 UTC m=+193.111512168 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.496563 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562700-h4fl7" Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.503547 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:01 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:01 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:01 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.503617 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.602197 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:01 crc kubenswrapper[4767]: E0317 15:40:01.609433 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:02.109404674 +0000 UTC m=+193.522720721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.609749 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:01 crc kubenswrapper[4767]: E0317 15:40:01.610315 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:02.110303977 +0000 UTC m=+193.523620024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:01 crc kubenswrapper[4767]: I0317 15:40:01.692309 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" event={"ID":"d3064e2c-7080-4395-9a85-8d8794d588e0","Type":"ContainerStarted","Data":"885dade4461052c804560ad31d3ebfaaf6ce33d9a529dae8417c1ba08000db40"} Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.129594 4767 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-pv6ds container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.129732 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" podUID="ece0e760-026e-4de3-bb90-bb5117963f69" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.135780 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.135828 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.135899 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.135917 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.146707 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:02 crc kubenswrapper[4767]: E0317 15:40:02.148751 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:03.148729668 +0000 UTC m=+194.562045715 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.699880 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:02 crc kubenswrapper[4767]: E0317 15:40:02.700891 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:03.20087341 +0000 UTC m=+194.614189457 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.724644 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:02 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:02 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:02 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.724718 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.736485 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" event={"ID":"b836be84-e8b0-4e8d-8bac-cfc922bea7eb","Type":"ContainerStarted","Data":"17d8403e3f262fd23c063dcbbb20f9f6124c19c1851c67e125a6ab143e33581f"} Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.737708 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.899102 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.899151 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 17 15:40:02 crc kubenswrapper[4767]: I0317 15:40:02.900212 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:02 crc kubenswrapper[4767]: E0317 15:40:02.901236 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:03.40121653 +0000 UTC m=+194.814532647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.008226 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:03 crc kubenswrapper[4767]: E0317 15:40:03.008798 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:03.508772135 +0000 UTC m=+194.922088182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.010647 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:03 crc kubenswrapper[4767]: E0317 15:40:03.011075 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:03.511062854 +0000 UTC m=+194.924378901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.110443 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" event={"ID":"2047989e-3a86-4a4e-8b4a-c8d29fc17e34","Type":"ContainerStarted","Data":"e136653f29e71e90b35e758962105cecb19706d4ee1b0e68ad34d7ea86a35bfa"} Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.113261 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:03 crc kubenswrapper[4767]: E0317 15:40:03.114629 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:03.614610266 +0000 UTC m=+195.027926313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.115898 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" event={"ID":"1c88a6ee-cd24-4d85-8a89-d830e5baa434","Type":"ContainerStarted","Data":"f9f55ad8c367bf4988ef6992df61b4226db9892e582b799b28991e35e810cbd0"} Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.117621 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w" event={"ID":"246c8a34-7be4-4e94-ab60-c4e471a762b1","Type":"ContainerStarted","Data":"9b7648cbb08d9bf2e1522d3eb574c96cc1aaa7241ca00c0e171941eebff60549"} Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.120886 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" event={"ID":"4c4aa56e-7fca-4836-b8ab-581ba3be753c","Type":"ContainerStarted","Data":"88c0f6a7910761f17a26e2a32933d2ffe7e98b3b262a3ff9220c0a33c9cdbe57"} Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.122604 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" event={"ID":"bbbc3925-9201-47fa-a29f-e6b271de61fd","Type":"ContainerStarted","Data":"c766d2ce56b4c431277777a116c7840206d8979ffd11ee660303f4f6127b933b"} Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.126015 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" event={"ID":"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc","Type":"ContainerStarted","Data":"cb8c86b937458536705cdf5e032ded29e32be4bf131efaabc33199b39b638764"} Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.128942 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" event={"ID":"3398275a-5195-45d6-9444-9e4d9825770f","Type":"ContainerStarted","Data":"3793d071d44367af6e92812e4bb654c1af619b6187460bbaf468855a9dc3d678"} Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.133452 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cspmp" event={"ID":"49254887-6adf-47f8-ab3b-8c8890281ee6","Type":"ContainerStarted","Data":"023f3157c33b9df1d7449863291763f9ad5d8bac670fad69af524bd99dc08631"} Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.135815 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" event={"ID":"68f6ff1d-a119-4faf-89aa-fd9afc2d93f8","Type":"ContainerStarted","Data":"4965483e6277ff80bc871b75394d2f20f37e0b634fd01578030291d8d29bfe8a"} Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.136413 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.139297 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-sdx96" event={"ID":"02b1adc8-dbd3-4cfd-88d0-f3c2f95b5509","Type":"ContainerStarted","Data":"e8d9b0f9712eb635725e822aa7d1387b924131947fc25075469e42e49ea64b0b"} Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.301132 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.545648 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" event={"ID":"99e547df-6c91-4f2b-b8c8-5a9ddcb942b3","Type":"ContainerStarted","Data":"097c47349c95e39cd1653c1b8bc7eba9f987945ac2111bb88b346d21c7c6aca9"} Mar 17 15:40:03 crc kubenswrapper[4767]: E0317 15:40:03.545745 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:04.045726639 +0000 UTC m=+195.459042756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.653525 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" event={"ID":"3b01694a-f9a3-4f54-ae85-7bb50a00f5ca","Type":"ContainerStarted","Data":"c0bb6f53faa9c3a57caa5c17b15aee78030fb0076ed54c1b01c8006e903e0690"} Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.654025 4767 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cb6sk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.654060 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" podUID="acdeabaa-85fb-4d85-bad6-89fc1be0125a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.654384 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:03 crc kubenswrapper[4767]: E0317 15:40:03.655421 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:04.155389957 +0000 UTC m=+195.568706004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.756998 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:03 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:03 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:03 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.757054 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:03 crc kubenswrapper[4767]: I0317 15:40:03.757874 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:03 crc kubenswrapper[4767]: E0317 15:40:03.758386 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:04.258374365 +0000 UTC m=+195.671690412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:04 crc kubenswrapper[4767]: I0317 15:40:03.933281 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:04 crc kubenswrapper[4767]: E0317 15:40:03.933908 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:04.43387112 +0000 UTC m=+195.847187167 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:04 crc kubenswrapper[4767]: I0317 15:40:03.934254 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:04 crc kubenswrapper[4767]: E0317 15:40:03.935916 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:04.435901372 +0000 UTC m=+195.849217479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:04 crc kubenswrapper[4767]: I0317 15:40:04.046299 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:04 crc kubenswrapper[4767]: E0317 15:40:04.046610 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:04.546595167 +0000 UTC m=+195.959911214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:04 crc kubenswrapper[4767]: I0317 15:40:04.179017 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:04 crc kubenswrapper[4767]: E0317 15:40:04.179748 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:04.679734288 +0000 UTC m=+196.093050335 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:04 crc kubenswrapper[4767]: I0317 15:40:04.457812 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:04 crc kubenswrapper[4767]: E0317 15:40:04.458983 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:04.958957459 +0000 UTC m=+196.372273506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:04 crc kubenswrapper[4767]: I0317 15:40:04.474911 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:04 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:04 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:04 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:04 crc kubenswrapper[4767]: I0317 15:40:04.474979 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:04.792782 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:04.793193 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:05.293161972 +0000 UTC m=+196.706478019 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:04.942836 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:04.943352 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:05.443336096 +0000 UTC m=+196.856652143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:05.103137 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:05.103785 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:05.603773415 +0000 UTC m=+197.017089462 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:05.501606 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:05.507483 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:06.007451435 +0000 UTC m=+197.420767482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:05.508350 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:11 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:05.508411 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.138346 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: i/o timeout" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.138408 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: i/o timeout" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.141081 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.141115 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.156018 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:06.195092 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:07.195035136 +0000 UTC m=+198.608351273 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.218782 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-cspmp" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.218852 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.218869 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cspmp" event={"ID":"49254887-6adf-47f8-ab3b-8c8890281ee6","Type":"ContainerStarted","Data":"53d7b43c008bd8aefa75051fed1da9db66676fa8cbba1b73a4354c69c1bfc5c1"} Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.218905 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.220506 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.220675 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.713367 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.717098 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:06.718259 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:07.218060348 +0000 UTC m=+198.631376395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.719591 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.719664 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.719715 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.720591 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:11 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.720616 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.723953 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.730056 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.731473 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.747716 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:06.748103 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:07.635254 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" event={"ID":"55851106-62aa-48a2-bd76-e0049671b476","Type":"ContainerStarted","Data":"f282815430bd823ec035a83e8483e9ba0ccb0d515d728f23cdf27f2e92bffb6f"} Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.270322 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:08.271562 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:09.271541419 +0000 UTC m=+200.684857466 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.272572 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ba34e593-ec5d-4e7e-9505-e21f3242f3b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.272626 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ba34e593-ec5d-4e7e-9505-e21f3242f3b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.276618 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.276665 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.294230 4767 patch_prober.go:28] interesting pod/console-f9d7485db-9cz8q container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.294251 4767 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cb6sk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.294291 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9cz8q" podUID="deba5665-b026-4d87-a986-b17f02f644d7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.294319 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" podUID="acdeabaa-85fb-4d85-bad6-89fc1be0125a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.294401 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.294420 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.294516 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="openshift-config-operator" containerStatusID={"Type":"cri-o","ID":"0f8553bfc41dec31090eb663209a1e4657345284d2f6241ad40d2b406f04d846"} pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" containerMessage="Container openshift-config-operator failed liveness probe, will be restarted" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.294570 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" containerID="cri-o://0f8553bfc41dec31090eb663209a1e4657345284d2f6241ad40d2b406f04d846" gracePeriod=30 Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.294877 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.294899 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.301565 4767 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-pv6ds container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.301627 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" podUID="ece0e760-026e-4de3-bb90-bb5117963f69" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.301736 4767 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rtlwt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.301763 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" podUID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.316919 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:11 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.316999 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.317545 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.317625 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.333346 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.333391 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.393102 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.393163 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.410484 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.410637 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97a26755-5343-4d6b-92b1-a05b5542944f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"97a26755-5343-4d6b-92b1-a05b5542944f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.410736 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97a26755-5343-4d6b-92b1-a05b5542944f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"97a26755-5343-4d6b-92b1-a05b5542944f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.410806 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ba34e593-ec5d-4e7e-9505-e21f3242f3b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.410822 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ba34e593-ec5d-4e7e-9505-e21f3242f3b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:08.411278 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:08.911249145 +0000 UTC m=+200.324565192 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.412442 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ba34e593-ec5d-4e7e-9505-e21f3242f3b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.554018 4767 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-tz4lw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.554149 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" podUID="a90441f8-30d7-4091-a23f-2522c586f8cf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.842128 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97a26755-5343-4d6b-92b1-a05b5542944f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"97a26755-5343-4d6b-92b1-a05b5542944f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.842228 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.842294 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97a26755-5343-4d6b-92b1-a05b5542944f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"97a26755-5343-4d6b-92b1-a05b5542944f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.842719 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97a26755-5343-4d6b-92b1-a05b5542944f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"97a26755-5343-4d6b-92b1-a05b5542944f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.842920 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:11 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.843040 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:08.844032 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:09.34382154 +0000 UTC m=+200.757137597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:08.952617 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:08.953398 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:09.453380939 +0000 UTC m=+200.866696986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.137616 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:09.246680 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:09.746666108 +0000 UTC m=+201.159982155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.253909 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.253976 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.354668 4767 patch_prober.go:28] interesting pod/dns-default-cspmp container/dns namespace/openshift-dns: Readiness probe status=failure output="Get \"http://10.217.0.39:8181/ready\": dial tcp 10.217.0.39:8181: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.354732 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-dns/dns-default-cspmp" podUID="49254887-6adf-47f8-ab3b-8c8890281ee6" containerName="dns" probeResult="failure" output="Get \"http://10.217.0.39:8181/ready\": dial tcp 10.217.0.39:8181: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.356221 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:09.359285 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:09.859255781 +0000 UTC m=+201.272571828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.384392 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:11 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.384449 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.568396 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:09.568916 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:10.068835785 +0000 UTC m=+201.482151832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.581329 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.581383 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.581394 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.581443 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.581864 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.581894 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.581903 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.581926 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.594515 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97a26755-5343-4d6b-92b1-a05b5542944f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"97a26755-5343-4d6b-92b1-a05b5542944f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.702408 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:09.702949 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:10.202927619 +0000 UTC m=+201.616243666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.704477 4767 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cb6sk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.704529 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" podUID="acdeabaa-85fb-4d85-bad6-89fc1be0125a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.704607 4767 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cb6sk container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.704624 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" podUID="acdeabaa-85fb-4d85-bad6-89fc1be0125a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:09.827316 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:09.827709 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:10.327690853 +0000 UTC m=+201.741006900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.025256 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:10.030244 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:10.530206468 +0000 UTC m=+201.943522525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.127482 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.127504 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.127552 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.127559 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.139199 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:10.139716 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:10.639701242 +0000 UTC m=+202.053017289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.244339 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:10.245004 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:10.744983146 +0000 UTC m=+202.158299193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.264751 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-config-operator_openshift-config-operator-7777fb866f-rphbj_3f692e30-f8c9-4823-bc82-7bf8abca0659/openshift-config-operator/0.log" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.318261 4767 generic.go:334] "Generic (PLEG): container finished" podID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerID="0f8553bfc41dec31090eb663209a1e4657345284d2f6241ad40d2b406f04d846" exitCode=2 Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.326404 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ba34e593-ec5d-4e7e-9505-e21f3242f3b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.335604 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.335869 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.336192 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.336211 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.345796 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:10.346243 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:10.846232168 +0000 UTC m=+202.259548215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:10.903829 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:10.904120 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:11.904095625 +0000 UTC m=+203.317411672 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.001837 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:11 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.001902 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.022991 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:11.023503 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:11.523486359 +0000 UTC m=+202.936802396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.127741 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.129065 4767 patch_prober.go:28] interesting pod/dns-default-cspmp container/dns namespace/openshift-dns: Readiness probe status=failure output="Get \"http://10.217.0.39:8181/ready\": dial tcp 10.217.0.39:8181: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.129101 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-dns/dns-default-cspmp" podUID="49254887-6adf-47f8-ab3b-8c8890281ee6" containerName="dns" probeResult="failure" output="Get \"http://10.217.0.39:8181/ready\": dial tcp 10.217.0.39:8181: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:11.130969 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:11.630946161 +0000 UTC m=+203.044262208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.226714 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.226763 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.228651 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:11.229315 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:11.729301241 +0000 UTC m=+203.142617288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.327379 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:11 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:11 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.327522 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.336784 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:11.337497 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:11.837475418 +0000 UTC m=+203.250791465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.528784 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:11.529391 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:12.029365957 +0000 UTC m=+203.442682004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.944237 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:11.944748 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:12.444727042 +0000 UTC m=+203.858043089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.966626 4767 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-pv6ds container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.978029 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" podUID="ece0e760-026e-4de3-bb90-bb5117963f69" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 17 15:40:11 crc kubenswrapper[4767]: E0317 15:40:11.984353 4767 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.349s" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.984454 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:40:11 crc kubenswrapper[4767]: I0317 15:40:11.984499 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" event={"ID":"3f692e30-f8c9-4823-bc82-7bf8abca0659","Type":"ContainerDied","Data":"0f8553bfc41dec31090eb663209a1e4657345284d2f6241ad40d2b406f04d846"} Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.083479 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:12 crc kubenswrapper[4767]: E0317 15:40:12.085301 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:12.585276094 +0000 UTC m=+203.998592141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.177391 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.186101 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:12 crc kubenswrapper[4767]: E0317 15:40:12.244030 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:12.743966539 +0000 UTC m=+204.157282606 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.269133 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:12 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:12 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:12 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.269498 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.287715 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:12 crc kubenswrapper[4767]: E0317 15:40:12.291320 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:12.791298653 +0000 UTC m=+204.204614710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.294119 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.389044 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:12 crc kubenswrapper[4767]: E0317 15:40:12.389436 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:12.889422375 +0000 UTC m=+204.302738422 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.446613 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.529301 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:12 crc kubenswrapper[4767]: E0317 15:40:12.530630 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:13.03060779 +0000 UTC m=+204.443923837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.543879 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" podStartSLOduration=138.543858699 podStartE2EDuration="2m18.543858699s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:12.542423538 +0000 UTC m=+203.955739595" watchObservedRunningTime="2026-03-17 15:40:12.543858699 +0000 UTC m=+203.957174746" Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.546662 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jhtfz" podStartSLOduration=137.546640807 podStartE2EDuration="2m17.546640807s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:12.440612446 +0000 UTC m=+203.853928493" watchObservedRunningTime="2026-03-17 15:40:12.546640807 +0000 UTC m=+203.959956854" Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.641415 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:12 crc kubenswrapper[4767]: E0317 15:40:12.642981 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:13.142948785 +0000 UTC m=+204.556264832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:12 crc kubenswrapper[4767]: I0317 15:40:12.899481 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:12 crc kubenswrapper[4767]: E0317 15:40:12.899832 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:13.399815082 +0000 UTC m=+204.813131139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.082128 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:13 crc kubenswrapper[4767]: E0317 15:40:13.082745 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:13.582727724 +0000 UTC m=+204.996043771 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.186996 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:13 crc kubenswrapper[4767]: E0317 15:40:13.187918 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:13.687896344 +0000 UTC m=+205.101212391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.208977 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" event={"ID":"bbbc3925-9201-47fa-a29f-e6b271de61fd","Type":"ContainerStarted","Data":"b806b507bc1c75932c96afc7f714cbf7a3787021b32eef573d046c63f7ac18b8"} Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.368009 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:13 crc kubenswrapper[4767]: E0317 15:40:13.370253 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:13.870221444 +0000 UTC m=+205.283537551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.405356 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:13 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:13 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:13 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.405431 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.421665 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4p5wd" podStartSLOduration=139.421637203 podStartE2EDuration="2m19.421637203s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:13.410946065 +0000 UTC m=+204.824262132" watchObservedRunningTime="2026-03-17 15:40:13.421637203 +0000 UTC m=+204.834953260" Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.424957 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podStartSLOduration=138.42493447 podStartE2EDuration="2m18.42493447s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:13.202718908 +0000 UTC m=+204.616034975" watchObservedRunningTime="2026-03-17 15:40:13.42493447 +0000 UTC m=+204.838250527" Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.453540 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" event={"ID":"2047989e-3a86-4a4e-8b4a-c8d29fc17e34","Type":"ContainerStarted","Data":"1f699325d8e96bf9a6171333e52c42f90f38a00f43e410145d5bbb02a69ad5e4"} Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.456198 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" event={"ID":"1c88a6ee-cd24-4d85-8a89-d830e5baa434","Type":"ContainerStarted","Data":"07b473ca20a6f8879b6f05593d0a8240dac4efae2906473b9b7897b1e634b375"} Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.492053 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" podStartSLOduration=138.492034934 podStartE2EDuration="2m18.492034934s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:13.491844857 +0000 UTC m=+204.905160914" watchObservedRunningTime="2026-03-17 15:40:13.492034934 +0000 UTC m=+204.905350981" Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.492357 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:13 crc kubenswrapper[4767]: E0317 15:40:13.492474 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:13.992452879 +0000 UTC m=+205.405768936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.549136 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:13 crc kubenswrapper[4767]: E0317 15:40:13.554969 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:14.05494862 +0000 UTC m=+205.468264667 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.559935 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562700-h4fl7"] Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.572302 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" event={"ID":"3b01694a-f9a3-4f54-ae85-7bb50a00f5ca","Type":"ContainerStarted","Data":"044ccbb030d841eb7513347876953b1479e11f8835efaca4b03c839e45f63650"} Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.573737 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-sdx96" podStartSLOduration=27.573723844 podStartE2EDuration="27.573723844s" podCreationTimestamp="2026-03-17 15:39:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:13.573132933 +0000 UTC m=+204.986448980" watchObservedRunningTime="2026-03-17 15:40:13.573723844 +0000 UTC m=+204.987039891" Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.603029 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podStartSLOduration=138.60301142 podStartE2EDuration="2m18.60301142s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:13.600007794 +0000 UTC m=+205.013323851" watchObservedRunningTime="2026-03-17 15:40:13.60301142 +0000 UTC m=+205.016327467" Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.624584 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podStartSLOduration=138.624566663 podStartE2EDuration="2m18.624566663s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:13.619541855 +0000 UTC m=+205.032857902" watchObservedRunningTime="2026-03-17 15:40:13.624566663 +0000 UTC m=+205.037882710" Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.652890 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:13 crc kubenswrapper[4767]: E0317 15:40:13.653238 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:14.153219996 +0000 UTC m=+205.566536043 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:13 crc kubenswrapper[4767]: I0317 15:40:13.917659 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:13 crc kubenswrapper[4767]: E0317 15:40:13.918754 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:14.41872491 +0000 UTC m=+205.832040967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.021858 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:14 crc kubenswrapper[4767]: E0317 15:40:14.022325 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:14.522299794 +0000 UTC m=+205.935615841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.057724 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.057780 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.118805 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.124333 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:14 crc kubenswrapper[4767]: E0317 15:40:14.124694 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:14.624677996 +0000 UTC m=+206.037994043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.133436 4767 patch_prober.go:28] interesting pod/dns-default-cspmp container/dns namespace/openshift-dns: Readiness probe status=failure output="Get \"http://10.217.0.39:8181/ready\": dial tcp 10.217.0.39:8181: connect: connection refused" start-of-body= Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.133700 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-dns/dns-default-cspmp" podUID="49254887-6adf-47f8-ab3b-8c8890281ee6" containerName="dns" probeResult="failure" output="Get \"http://10.217.0.39:8181/ready\": dial tcp 10.217.0.39:8181: connect: connection refused" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.226436 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:14 crc kubenswrapper[4767]: E0317 15:40:14.237292 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:14.737263999 +0000 UTC m=+206.150580046 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.278737 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2v6ms" podStartSLOduration=139.278708526 podStartE2EDuration="2m19.278708526s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:14.198735186 +0000 UTC m=+205.612051253" watchObservedRunningTime="2026-03-17 15:40:14.278708526 +0000 UTC m=+205.692024573" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.307427 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rtlwt"] Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.307922 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" podUID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" containerName="controller-manager" containerID="cri-o://a9715511abf3e82c7c36b395096eec861e4c5f36b40fa5921a952218d7781141" gracePeriod=30 Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.308747 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:14 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:14 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:14 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.309014 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.319528 4767 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rtlwt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.319780 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" podUID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.329319 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:14 crc kubenswrapper[4767]: E0317 15:40:14.329915 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:14.829899867 +0000 UTC m=+206.243215914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.368265 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-cspmp" podStartSLOduration=28.368247824 podStartE2EDuration="28.368247824s" podCreationTimestamp="2026-03-17 15:39:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:14.350453544 +0000 UTC m=+205.763769591" watchObservedRunningTime="2026-03-17 15:40:14.368247824 +0000 UTC m=+205.781563871" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.386765 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw"] Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.386985 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" podUID="a90441f8-30d7-4091-a23f-2522c586f8cf" containerName="route-controller-manager" containerID="cri-o://3efe3ea49c08258f3ba548a8ea7c07dce8d7dd6c72099fdb81abb6d455a891a9" gracePeriod=30 Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.410490 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.431728 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:14 crc kubenswrapper[4767]: E0317 15:40:14.432090 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:14.932073702 +0000 UTC m=+206.345389749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.436749 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4twt" podStartSLOduration=140.436726236 podStartE2EDuration="2m20.436726236s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:14.410144556 +0000 UTC m=+205.823460613" watchObservedRunningTime="2026-03-17 15:40:14.436726236 +0000 UTC m=+205.850042283" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.531431 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-vwszx" podStartSLOduration=139.531409256 podStartE2EDuration="2m19.531409256s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:14.482273158 +0000 UTC m=+205.895589205" watchObservedRunningTime="2026-03-17 15:40:14.531409256 +0000 UTC m=+205.944725303" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.535999 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:14 crc kubenswrapper[4767]: E0317 15:40:14.536476 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:15.036463425 +0000 UTC m=+206.449779472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.543288 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.543669 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.566026 4767 patch_prober.go:28] interesting pod/apiserver-76f77b778f-bk7p5 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.36:8443/livez\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.566087 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" podUID="1c88a6ee-cd24-4d85-8a89-d830e5baa434" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.36:8443/livez\": dial tcp 10.217.0.36:8443: connect: connection refused" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.639791 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:14 crc kubenswrapper[4767]: E0317 15:40:14.640043 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:15.140003498 +0000 UTC m=+206.553319545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.640233 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:14 crc kubenswrapper[4767]: E0317 15:40:14.640523 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:15.140511616 +0000 UTC m=+206.553827653 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.671837 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-879f6c89f-rtlwt_8a0ddaae-cb65-45e0-a82a-44f6afb10d48/controller-manager/0.log" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.672043 4767 generic.go:334] "Generic (PLEG): container finished" podID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" containerID="a9715511abf3e82c7c36b395096eec861e4c5f36b40fa5921a952218d7781141" exitCode=2 Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.672116 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" event={"ID":"8a0ddaae-cb65-45e0-a82a-44f6afb10d48","Type":"ContainerDied","Data":"a9715511abf3e82c7c36b395096eec861e4c5f36b40fa5921a952218d7781141"} Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.690586 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" event={"ID":"3b01694a-f9a3-4f54-ae85-7bb50a00f5ca","Type":"ContainerStarted","Data":"ba9fc02a160a3816aab2118e7d342f4dbad8e5e2fd14f56ad856862c5674ff66"} Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.702036 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-frl2d" event={"ID":"201932aa-c539-434b-992e-92f889e52de3","Type":"ContainerStarted","Data":"b3c845649071c417479a4245863ffaa34118c25e4da74b2155cd09a2de0c1ae2"} Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.714824 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t7mfn" event={"ID":"32925427-1529-4e8d-a606-8ccbde1fd877","Type":"ContainerStarted","Data":"3df17bc34dc96397766c1daf167de08e1f1b7f6e9a7f6077a35fdb474c06ecb3"} Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.942369 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:14 crc kubenswrapper[4767]: E0317 15:40:14.943624 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:15.443599339 +0000 UTC m=+206.856915436 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.951258 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-config-operator_openshift-config-operator-7777fb866f-rphbj_3f692e30-f8c9-4823-bc82-7bf8abca0659/openshift-config-operator/0.log" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.957520 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" event={"ID":"3f692e30-f8c9-4823-bc82-7bf8abca0659","Type":"ContainerStarted","Data":"d1e085d042d9d8b5cec9ab40cfe208b9224e0fe81e8ca191bf191605cd7af4c7"} Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.958329 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.968644 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pbr4w" podStartSLOduration=139.968625365 podStartE2EDuration="2m19.968625365s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:14.552401869 +0000 UTC m=+205.965717926" watchObservedRunningTime="2026-03-17 15:40:14.968625365 +0000 UTC m=+206.381941412" Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.969712 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562700-h4fl7" event={"ID":"051f576f-ce90-4401-8944-a340dd67b274","Type":"ContainerStarted","Data":"d57d73aa046e2e0fbbb9d78f1e72bcf4005f9031f7daf199e4c4411a2828f5fa"} Mar 17 15:40:14 crc kubenswrapper[4767]: I0317 15:40:14.971560 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26l9f" podStartSLOduration=140.971540698 podStartE2EDuration="2m20.971540698s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:14.967971821 +0000 UTC m=+206.381287888" watchObservedRunningTime="2026-03-17 15:40:14.971540698 +0000 UTC m=+206.384856745" Mar 17 15:40:15 crc kubenswrapper[4767]: I0317 15:40:15.078878 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:15 crc kubenswrapper[4767]: E0317 15:40:15.087271 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:15.587250851 +0000 UTC m=+207.000566898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:15 crc kubenswrapper[4767]: I0317 15:40:15.181406 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:15 crc kubenswrapper[4767]: E0317 15:40:15.182718 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:15.682700428 +0000 UTC m=+207.096016475 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:15 crc kubenswrapper[4767]: I0317 15:40:15.286405 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:15 crc kubenswrapper[4767]: E0317 15:40:15.286908 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:15.786887764 +0000 UTC m=+207.200203811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:15 crc kubenswrapper[4767]: I0317 15:40:15.290815 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:15 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:15 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:15 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:15 crc kubenswrapper[4767]: I0317 15:40:15.290889 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:15 crc kubenswrapper[4767]: I0317 15:40:15.388893 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:15 crc kubenswrapper[4767]: E0317 15:40:15.389372 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:15.88935494 +0000 UTC m=+207.302670987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:15 crc kubenswrapper[4767]: I0317 15:40:15.511071 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:15 crc kubenswrapper[4767]: E0317 15:40:15.528398 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:16.014879501 +0000 UTC m=+207.428195548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:15 crc kubenswrapper[4767]: I0317 15:40:15.618748 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:15 crc kubenswrapper[4767]: E0317 15:40:15.619288 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:16.119265823 +0000 UTC m=+207.532581870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:15 crc kubenswrapper[4767]: I0317 15:40:15.755946 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:15 crc kubenswrapper[4767]: E0317 15:40:15.756350 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:16.256338523 +0000 UTC m=+207.669654570 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:15 crc kubenswrapper[4767]: I0317 15:40:15.858036 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:15 crc kubenswrapper[4767]: E0317 15:40:15.858548 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:16.358526528 +0000 UTC m=+207.771842575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:15 crc kubenswrapper[4767]: I0317 15:40:15.974860 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:15 crc kubenswrapper[4767]: E0317 15:40:15.975741 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:16.475723065 +0000 UTC m=+207.889039112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.057846 4767 generic.go:334] "Generic (PLEG): container finished" podID="a90441f8-30d7-4091-a23f-2522c586f8cf" containerID="3efe3ea49c08258f3ba548a8ea7c07dce8d7dd6c72099fdb81abb6d455a891a9" exitCode=0 Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.059092 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" event={"ID":"a90441f8-30d7-4091-a23f-2522c586f8cf","Type":"ContainerDied","Data":"3efe3ea49c08258f3ba548a8ea7c07dce8d7dd6c72099fdb81abb6d455a891a9"} Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.062450 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.062544 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.081372 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:16 crc kubenswrapper[4767]: E0317 15:40:16.081932 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:16.581912982 +0000 UTC m=+207.995229029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.189673 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:16 crc kubenswrapper[4767]: E0317 15:40:16.190634 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:16.690614458 +0000 UTC m=+208.103930555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.268244 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:16 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:16 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:16 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.268645 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.516642 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:16 crc kubenswrapper[4767]: E0317 15:40:16.517040 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:17.017021536 +0000 UTC m=+208.430337583 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.632464 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:16 crc kubenswrapper[4767]: E0317 15:40:16.632793 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:17.132779781 +0000 UTC m=+208.546095828 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.684974 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" podStartSLOduration=141.684940627 podStartE2EDuration="2m21.684940627s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:16.27461965 +0000 UTC m=+207.687935707" watchObservedRunningTime="2026-03-17 15:40:16.684940627 +0000 UTC m=+208.098256674" Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.686198 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.736887 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:16 crc kubenswrapper[4767]: E0317 15:40:16.737305 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:17.237287608 +0000 UTC m=+208.650603655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:16 crc kubenswrapper[4767]: W0317 15:40:16.751442 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podba34e593_ec5d_4e7e_9505_e21f3242f3b7.slice/crio-d6484187f76ec5fb1c4fde08f54929198dac176e5ebb78078768ffda550573ca WatchSource:0}: Error finding container d6484187f76ec5fb1c4fde08f54929198dac176e5ebb78078768ffda550573ca: Status 404 returned error can't find the container with id d6484187f76ec5fb1c4fde08f54929198dac176e5ebb78078768ffda550573ca Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.751501 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.921287 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:16 crc kubenswrapper[4767]: E0317 15:40:16.921704 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:17.421689801 +0000 UTC m=+208.835005848 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.931565 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2cwjx" podStartSLOduration=141.93154635 podStartE2EDuration="2m21.93154635s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:16.777092026 +0000 UTC m=+208.190408083" watchObservedRunningTime="2026-03-17 15:40:16.93154635 +0000 UTC m=+208.344862397" Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.931722 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-cwz8x" podStartSLOduration=141.931717566 podStartE2EDuration="2m21.931717566s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:16.931113225 +0000 UTC m=+208.344429272" watchObservedRunningTime="2026-03-17 15:40:16.931717566 +0000 UTC m=+208.345033613" Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.992056 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqjqr" podStartSLOduration=142.99203833 podStartE2EDuration="2m22.99203833s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:16.991588414 +0000 UTC m=+208.404904461" watchObservedRunningTime="2026-03-17 15:40:16.99203833 +0000 UTC m=+208.405354377" Mar 17 15:40:16 crc kubenswrapper[4767]: I0317 15:40:16.994279 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.023702 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:17 crc kubenswrapper[4767]: E0317 15:40:17.024132 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:17.524113955 +0000 UTC m=+208.937430002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.059404 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.059488 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.060074 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.060144 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.125106 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:17 crc kubenswrapper[4767]: E0317 15:40:17.148705 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:17.648681142 +0000 UTC m=+209.061997199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.199987 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-cspmp" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.227996 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:17 crc kubenswrapper[4767]: E0317 15:40:17.229234 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:17.729213711 +0000 UTC m=+209.142529758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.248406 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.267909 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"97a26755-5343-4d6b-92b1-a05b5542944f","Type":"ContainerStarted","Data":"6716b310725f103621b887bacafbf1e8cbaa7d078d4bde96b5561d70c270499a"} Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.270624 4767 generic.go:334] "Generic (PLEG): container finished" podID="2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc" containerID="cb8c86b937458536705cdf5e032ded29e32be4bf131efaabc33199b39b638764" exitCode=0 Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.270677 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" event={"ID":"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc","Type":"ContainerDied","Data":"cb8c86b937458536705cdf5e032ded29e32be4bf131efaabc33199b39b638764"} Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.272688 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ba34e593-ec5d-4e7e-9505-e21f3242f3b7","Type":"ContainerStarted","Data":"d6484187f76ec5fb1c4fde08f54929198dac176e5ebb78078768ffda550573ca"} Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.273087 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.273119 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.276876 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:17 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:17 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:17 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.276906 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.378496 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:17 crc kubenswrapper[4767]: E0317 15:40:17.379487 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:17.879467237 +0000 UTC m=+209.292783284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.468978 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-879f6c89f-rtlwt_8a0ddaae-cb65-45e0-a82a-44f6afb10d48/controller-manager/0.log" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.469060 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.520310 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-t7mfn" podStartSLOduration=142.520291639 podStartE2EDuration="2m22.520291639s" podCreationTimestamp="2026-03-17 15:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:17.499258355 +0000 UTC m=+208.912574422" watchObservedRunningTime="2026-03-17 15:40:17.520291639 +0000 UTC m=+208.933607686" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.564062 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.564117 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-client-ca\") pod \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.564142 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-proxy-ca-bundles\") pod \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.564195 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-config\") pod \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " Mar 17 15:40:17 crc kubenswrapper[4767]: E0317 15:40:17.564225 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:18.064208773 +0000 UTC m=+209.477524820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.564244 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-serving-cert\") pod \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.564355 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.564950 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-config" (OuterVolumeSpecName: "config") pod "8a0ddaae-cb65-45e0-a82a-44f6afb10d48" (UID: "8a0ddaae-cb65-45e0-a82a-44f6afb10d48"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.566467 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-client-ca" (OuterVolumeSpecName: "client-ca") pod "8a0ddaae-cb65-45e0-a82a-44f6afb10d48" (UID: "8a0ddaae-cb65-45e0-a82a-44f6afb10d48"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:40:17 crc kubenswrapper[4767]: E0317 15:40:17.566604 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:18.066594377 +0000 UTC m=+209.479910414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.566900 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8a0ddaae-cb65-45e0-a82a-44f6afb10d48" (UID: "8a0ddaae-cb65-45e0-a82a-44f6afb10d48"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.801716 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.801791 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbf4c\" (UniqueName: \"kubernetes.io/projected/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-kube-api-access-jbf4c\") pod \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\" (UID: \"8a0ddaae-cb65-45e0-a82a-44f6afb10d48\") " Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.801979 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.801989 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.801998 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:17 crc kubenswrapper[4767]: E0317 15:40:17.844877 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:18.344849112 +0000 UTC m=+209.758165159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.865927 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-kube-api-access-jbf4c" (OuterVolumeSpecName: "kube-api-access-jbf4c") pod "8a0ddaae-cb65-45e0-a82a-44f6afb10d48" (UID: "8a0ddaae-cb65-45e0-a82a-44f6afb10d48"). InnerVolumeSpecName "kube-api-access-jbf4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.875800 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8a0ddaae-cb65-45e0-a82a-44f6afb10d48" (UID: "8a0ddaae-cb65-45e0-a82a-44f6afb10d48"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.947395 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.947514 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.947528 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbf4c\" (UniqueName: \"kubernetes.io/projected/8a0ddaae-cb65-45e0-a82a-44f6afb10d48-kube-api-access-jbf4c\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:17 crc kubenswrapper[4767]: E0317 15:40:17.947759 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:18.447745052 +0000 UTC m=+209.861061099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:17 crc kubenswrapper[4767]: I0317 15:40:17.977713 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.033352 4767 patch_prober.go:28] interesting pod/console-f9d7485db-9cz8q container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.033409 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9cz8q" podUID="deba5665-b026-4d87-a986-b17f02f644d7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.047999 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:18 crc kubenswrapper[4767]: E0317 15:40:18.048399 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:18.548383492 +0000 UTC m=+209.961699539 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.172241 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-config\") pod \"a90441f8-30d7-4091-a23f-2522c586f8cf\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.172593 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4946\" (UniqueName: \"kubernetes.io/projected/a90441f8-30d7-4091-a23f-2522c586f8cf-kube-api-access-p4946\") pod \"a90441f8-30d7-4091-a23f-2522c586f8cf\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.172644 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-client-ca\") pod \"a90441f8-30d7-4091-a23f-2522c586f8cf\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.172671 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a90441f8-30d7-4091-a23f-2522c586f8cf-serving-cert\") pod \"a90441f8-30d7-4091-a23f-2522c586f8cf\" (UID: \"a90441f8-30d7-4091-a23f-2522c586f8cf\") " Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.172775 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.172804 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" podStartSLOduration=144.172786904 podStartE2EDuration="2m24.172786904s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:18.165145234 +0000 UTC m=+209.578461301" watchObservedRunningTime="2026-03-17 15:40:18.172786904 +0000 UTC m=+209.586102951" Mar 17 15:40:18 crc kubenswrapper[4767]: E0317 15:40:18.173212 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:18.673197649 +0000 UTC m=+210.086513696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.173840 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-client-ca" (OuterVolumeSpecName: "client-ca") pod "a90441f8-30d7-4091-a23f-2522c586f8cf" (UID: "a90441f8-30d7-4091-a23f-2522c586f8cf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.179750 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-config" (OuterVolumeSpecName: "config") pod "a90441f8-30d7-4091-a23f-2522c586f8cf" (UID: "a90441f8-30d7-4091-a23f-2522c586f8cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.197652 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90441f8-30d7-4091-a23f-2522c586f8cf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a90441f8-30d7-4091-a23f-2522c586f8cf" (UID: "a90441f8-30d7-4091-a23f-2522c586f8cf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.200276 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90441f8-30d7-4091-a23f-2522c586f8cf-kube-api-access-p4946" (OuterVolumeSpecName: "kube-api-access-p4946") pod "a90441f8-30d7-4091-a23f-2522c586f8cf" (UID: "a90441f8-30d7-4091-a23f-2522c586f8cf"). InnerVolumeSpecName "kube-api-access-p4946". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.341810 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.342125 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.342143 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4946\" (UniqueName: \"kubernetes.io/projected/a90441f8-30d7-4091-a23f-2522c586f8cf-kube-api-access-p4946\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.342155 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a90441f8-30d7-4091-a23f-2522c586f8cf-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.342164 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a90441f8-30d7-4091-a23f-2522c586f8cf-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:18 crc kubenswrapper[4767]: E0317 15:40:18.343623 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:18.843595587 +0000 UTC m=+210.256911654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.480199 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.480254 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.480344 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-59hsl" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.480350 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:18 crc kubenswrapper[4767]: E0317 15:40:18.480705 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:18.980691497 +0000 UTC m=+210.394007544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.480731 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"c82a98015caa0bb7c218c4c3adb06778f84051c358433937ea7560b971a25552"} pod="openshift-console/downloads-7954f5f757-59hsl" containerMessage="Container download-server failed liveness probe, will be restarted" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.480757 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" containerID="cri-o://c82a98015caa0bb7c218c4c3adb06778f84051c358433937ea7560b971a25552" gracePeriod=2 Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.480784 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.480805 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.481231 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.481277 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.482017 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:18 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:18 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:18 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.482042 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.582807 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:18 crc kubenswrapper[4767]: E0317 15:40:18.583424 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:19.083401291 +0000 UTC m=+210.496717338 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.584478 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-879f6c89f-rtlwt_8a0ddaae-cb65-45e0-a82a-44f6afb10d48/controller-manager/0.log" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.584549 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" event={"ID":"8a0ddaae-cb65-45e0-a82a-44f6afb10d48","Type":"ContainerDied","Data":"cd2135ec1df21d32437aca42407833fcaccc52595df61dc7fa36d5660bd66fa9"} Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.584584 4767 scope.go:117] "RemoveContainer" containerID="a9715511abf3e82c7c36b395096eec861e4c5f36b40fa5921a952218d7781141" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.584726 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rtlwt" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.596102 4767 ???:1] "http: TLS handshake error from 192.168.126.11:57362: no serving certificate available for the kubelet" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.661921 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.680958 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw" event={"ID":"a90441f8-30d7-4091-a23f-2522c586f8cf","Type":"ContainerDied","Data":"c883bbe7f75c93fb45ffd6f21675410a929bddce1f169375f941ba43c25f6251"} Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.683638 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:18 crc kubenswrapper[4767]: E0317 15:40:18.684066 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:19.184053422 +0000 UTC m=+210.597369469 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.765452 4767 scope.go:117] "RemoveContainer" containerID="3efe3ea49c08258f3ba548a8ea7c07dce8d7dd6c72099fdb81abb6d455a891a9" Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.784410 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:18 crc kubenswrapper[4767]: E0317 15:40:18.784825 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:19.284806527 +0000 UTC m=+210.698122574 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:18 crc kubenswrapper[4767]: I0317 15:40:18.886817 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:18 crc kubenswrapper[4767]: E0317 15:40:18.887391 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:19.387372935 +0000 UTC m=+210.800688982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.070503 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:19 crc kubenswrapper[4767]: E0317 15:40:19.070812 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:19.570793645 +0000 UTC m=+210.984109692 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.179691 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:19 crc kubenswrapper[4767]: E0317 15:40:19.180142 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:19.680125303 +0000 UTC m=+211.093441350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.269637 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:19 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:19 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:19 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.269717 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.292403 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:19 crc kubenswrapper[4767]: E0317 15:40:19.292701 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:19.792684805 +0000 UTC m=+211.206000852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.599957 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:19 crc kubenswrapper[4767]: E0317 15:40:19.600732 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:20.100717913 +0000 UTC m=+211.514033960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.619516 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.637806 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.746785 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:19 crc kubenswrapper[4767]: E0317 15:40:19.747629 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:20.24760269 +0000 UTC m=+211.660918797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.875004 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.875338 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:19 crc kubenswrapper[4767]: E0317 15:40:19.877024 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:20.377007127 +0000 UTC m=+211.790323254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.896133 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ba34e593-ec5d-4e7e-9505-e21f3242f3b7","Type":"ContainerStarted","Data":"853074c8dd907aad73979fb890b19677646d3dd46327a4c6c7bc585d1b71bd09"} Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.903448 4767 ???:1] "http: TLS handshake error from 192.168.126.11:57370: no serving certificate available for the kubelet" Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.928579 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"97a26755-5343-4d6b-92b1-a05b5542944f","Type":"ContainerStarted","Data":"faa4b40271082a42499613a1f75213c9735de220ba7b688fa9d162245af63f6c"} Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.960734 4767 generic.go:334] "Generic (PLEG): container finished" podID="471b6054-3b2a-47c9-8889-942603600da9" containerID="c82a98015caa0bb7c218c4c3adb06778f84051c358433937ea7560b971a25552" exitCode=0 Mar 17 15:40:19 crc kubenswrapper[4767]: I0317 15:40:19.960793 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-59hsl" event={"ID":"471b6054-3b2a-47c9-8889-942603600da9","Type":"ContainerDied","Data":"c82a98015caa0bb7c218c4c3adb06778f84051c358433937ea7560b971a25552"} Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.002449 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.026375 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:20.526356572 +0000 UTC m=+211.939672619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.132743 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.133206 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:20.633187991 +0000 UTC m=+212.046504038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.144376 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.149498 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-676bc56c59-88gvd"] Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.149724 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90441f8-30d7-4091-a23f-2522c586f8cf" containerName="route-controller-manager" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.149734 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90441f8-30d7-4091-a23f-2522c586f8cf" containerName="route-controller-manager" Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.149747 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" containerName="controller-manager" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.149753 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" containerName="controller-manager" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.149849 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" containerName="controller-manager" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.149861 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a90441f8-30d7-4091-a23f-2522c586f8cf" containerName="route-controller-manager" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.150242 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-676bc56c59-88gvd" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.152594 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-676bc56c59-88gvd"] Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.156901 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.157083 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.164472 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.164729 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.164885 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.167956 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.168794 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.171417 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-676bc56c59-88gvd"] Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.183510 4767 ???:1] "http: TLS handshake error from 192.168.126.11:57382: no serving certificate available for the kubelet" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.195661 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66dc66f96c-x2grn"] Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.204825 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.211209 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-c85wx proxy-ca-bundles serving-cert], unattached volumes=[], failed to process volumes=[client-ca config kube-api-access-c85wx proxy-ca-bundles serving-cert]: context canceled" pod="openshift-controller-manager/controller-manager-676bc56c59-88gvd" podUID="f7ffd372-5054-48ec-8195-dc89cf096f38" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.234710 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.235371 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:20.735355376 +0000 UTC m=+212.148671423 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.247136 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66dc66f96c-x2grn"] Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.270561 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:20 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:20 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:20 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.270620 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.278367 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs"] Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.279153 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.292917 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.293080 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.293187 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.293328 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.293443 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.300516 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.329777 4767 ???:1] "http: TLS handshake error from 192.168.126.11:57396: no serving certificate available for the kubelet" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.329863 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs"] Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.332668 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d2kvv"] Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.335808 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.335860 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-config\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.335879 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-serving-cert\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.335912 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f5b7\" (UniqueName: \"kubernetes.io/projected/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-kube-api-access-8f5b7\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.335932 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-config\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.335980 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3569d92-3c22-4a58-b535-0400b94005a2-serving-cert\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.335997 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-proxy-ca-bundles\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.336018 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-client-ca\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.336032 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gb22\" (UniqueName: \"kubernetes.io/projected/c3569d92-3c22-4a58-b535-0400b94005a2-kube-api-access-4gb22\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.336047 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-client-ca\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.336329 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:20.836318617 +0000 UTC m=+212.249634664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.367585 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=15.367560002 podStartE2EDuration="15.367560002s" podCreationTimestamp="2026-03-17 15:40:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:20.365888793 +0000 UTC m=+211.779204840" watchObservedRunningTime="2026-03-17 15:40:20.367560002 +0000 UTC m=+211.780876049" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.378846 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.384505 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.448156 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.448664 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-utilities\") pod \"certified-operators-d2kvv\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.448738 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lp5b\" (UniqueName: \"kubernetes.io/projected/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-kube-api-access-5lp5b\") pod \"certified-operators-d2kvv\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.448789 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3569d92-3c22-4a58-b535-0400b94005a2-serving-cert\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.458930 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-proxy-ca-bundles\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.458999 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gb22\" (UniqueName: \"kubernetes.io/projected/c3569d92-3c22-4a58-b535-0400b94005a2-kube-api-access-4gb22\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.459025 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-client-ca\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.459052 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-client-ca\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.459131 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-catalog-content\") pod \"certified-operators-d2kvv\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.459190 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-config\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.459217 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-serving-cert\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.459249 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f5b7\" (UniqueName: \"kubernetes.io/projected/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-kube-api-access-8f5b7\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.459291 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-config\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.491554 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d2kvv"] Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.501774 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-config\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.502807 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:21.002788096 +0000 UTC m=+212.416104143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.573299 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-client-ca\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.620089 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-client-ca\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.628770 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-config\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.629243 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3569d92-3c22-4a58-b535-0400b94005a2-serving-cert\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.629830 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.629866 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-catalog-content\") pod \"certified-operators-d2kvv\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.629956 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-utilities\") pod \"certified-operators-d2kvv\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.630013 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lp5b\" (UniqueName: \"kubernetes.io/projected/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-kube-api-access-5lp5b\") pod \"certified-operators-d2kvv\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.630303 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:21.130287117 +0000 UTC m=+212.543603164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.630753 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-catalog-content\") pod \"certified-operators-d2kvv\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.630941 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-utilities\") pod \"certified-operators-d2kvv\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.633990 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-serving-cert\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.689161 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.706466 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f5b7\" (UniqueName: \"kubernetes.io/projected/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-kube-api-access-8f5b7\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.721710 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lp5b\" (UniqueName: \"kubernetes.io/projected/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-kube-api-access-5lp5b\") pod \"certified-operators-d2kvv\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.724835 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gb22\" (UniqueName: \"kubernetes.io/projected/c3569d92-3c22-4a58-b535-0400b94005a2-kube-api-access-4gb22\") pod \"route-controller-manager-c545b76c8-srlrs\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.743601 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-config-volume\") pod \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.743649 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7756\" (UniqueName: \"kubernetes.io/projected/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-kube-api-access-r7756\") pod \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.743707 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-secret-volume\") pod \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\" (UID: \"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc\") " Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.743850 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.745268 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:21.245247895 +0000 UTC m=+212.658563942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.745296 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-config-volume" (OuterVolumeSpecName: "config-volume") pod "2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc" (UID: "2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.773089 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.773413 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.774072 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:21.274056454 +0000 UTC m=+212.687372501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.774279 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc" (UID: "2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.775444 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-proxy-ca-bundles\") pod \"controller-manager-66dc66f96c-x2grn\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.777628 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-kube-api-access-r7756" (OuterVolumeSpecName: "kube-api-access-r7756") pod "2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc" (UID: "2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc"). InnerVolumeSpecName "kube-api-access-r7756". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.777897 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-klkzj"] Mar 17 15:40:20 crc kubenswrapper[4767]: E0317 15:40:20.778275 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc" containerName="collect-profiles" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.778292 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc" containerName="collect-profiles" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.778716 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc" containerName="collect-profiles" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.791421 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.794041 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.794404 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.794859 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-klkzj"] Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.803477 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-79fk6"] Mar 17 15:40:20 crc kubenswrapper[4767]: I0317 15:40:20.808871 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:20.992376 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-frl2d" event={"ID":"201932aa-c539-434b-992e-92f889e52de3","Type":"ContainerStarted","Data":"c46d6edc9e088fe5b8013a59073e3ca82f88f1a439c75adddb781c832f33d4cd"} Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:20.994946 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-59hsl" event={"ID":"471b6054-3b2a-47c9-8889-942603600da9","Type":"ContainerStarted","Data":"140aaa6eab7b06985ec3865ee30e1ab9ae5685144782d645d3c774697c868c33"} Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:20.996566 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-59hsl" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.002341 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-676bc56c59-88gvd" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.003864 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.003974 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f" event={"ID":"2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc","Type":"ContainerDied","Data":"faed4d92fa5b4a362d0f03b11d99bc741fc8dd24fc219ac197f3b20554f63dbd"} Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.004004 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="faed4d92fa5b4a362d0f03b11d99bc741fc8dd24fc219ac197f3b20554f63dbd" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.061208 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.061269 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.061398 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.061278 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.159708 4767 ???:1] "http: TLS handshake error from 192.168.126.11:35790: no serving certificate available for the kubelet" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.159833 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.159879 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.160454 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.161549 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.162022 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-catalog-content\") pod \"community-operators-klkzj\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.162043 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8lw6\" (UniqueName: \"kubernetes.io/projected/3dc430ad-a22a-4fe8-a4dd-5c94622da680-kube-api-access-r8lw6\") pod \"community-operators-klkzj\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.162060 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-utilities\") pod \"community-operators-klkzj\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.162341 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7756\" (UniqueName: \"kubernetes.io/projected/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-kube-api-access-r7756\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.162416 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.162454 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:21 crc kubenswrapper[4767]: E0317 15:40:21.162604 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:21.66259021 +0000 UTC m=+213.075906257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.180812 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-79fk6"] Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.195336 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-676bc56c59-88gvd" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.203233 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rq699"] Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.204355 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rq699"] Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.204447 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rq699" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.209051 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rtlwt"] Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.213808 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rtlwt"] Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.217593 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw"] Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.218806 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tz4lw"] Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.222797 4767 ???:1] "http: TLS handshake error from 192.168.126.11:35800: no serving certificate available for the kubelet" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.263787 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-catalog-content\") pod \"certified-operators-79fk6\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.263842 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhxhp\" (UniqueName: \"kubernetes.io/projected/5df92c46-85d1-4900-9916-486aa896d42a-kube-api-access-zhxhp\") pod \"certified-operators-79fk6\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.263914 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-catalog-content\") pod \"community-operators-klkzj\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.263955 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8lw6\" (UniqueName: \"kubernetes.io/projected/3dc430ad-a22a-4fe8-a4dd-5c94622da680-kube-api-access-r8lw6\") pod \"community-operators-klkzj\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.263989 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-utilities\") pod \"community-operators-klkzj\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.264096 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-utilities\") pod \"certified-operators-79fk6\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.264195 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:21 crc kubenswrapper[4767]: E0317 15:40:21.264729 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:21.764714683 +0000 UTC m=+213.178030740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.265950 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-catalog-content\") pod \"community-operators-klkzj\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.266785 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-utilities\") pod \"community-operators-klkzj\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.386593 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:21 crc kubenswrapper[4767]: E0317 15:40:21.386759 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:21.88674026 +0000 UTC m=+213.300056307 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.386796 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.386831 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhxhp\" (UniqueName: \"kubernetes.io/projected/5df92c46-85d1-4900-9916-486aa896d42a-kube-api-access-zhxhp\") pod \"certified-operators-79fk6\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.386852 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-catalog-content\") pod \"certified-operators-79fk6\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.386880 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-catalog-content\") pod \"community-operators-rq699\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " pod="openshift-marketplace/community-operators-rq699" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.386913 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-utilities\") pod \"community-operators-rq699\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " pod="openshift-marketplace/community-operators-rq699" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.386938 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nm8p\" (UniqueName: \"kubernetes.io/projected/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-kube-api-access-6nm8p\") pod \"community-operators-rq699\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " pod="openshift-marketplace/community-operators-rq699" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.387008 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-utilities\") pod \"certified-operators-79fk6\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:40:21 crc kubenswrapper[4767]: E0317 15:40:21.387088 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:21.887080412 +0000 UTC m=+213.300396459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.387390 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-utilities\") pod \"certified-operators-79fk6\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.387418 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-catalog-content\") pod \"certified-operators-79fk6\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.400721 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=17.400700824 podStartE2EDuration="17.400700824s" podCreationTimestamp="2026-03-17 15:40:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:21.39945615 +0000 UTC m=+212.812772217" watchObservedRunningTime="2026-03-17 15:40:21.400700824 +0000 UTC m=+212.814016871" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.408246 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:21 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:21 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:21 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.408295 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.589717 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhxhp\" (UniqueName: \"kubernetes.io/projected/5df92c46-85d1-4900-9916-486aa896d42a-kube-api-access-zhxhp\") pod \"certified-operators-79fk6\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.707664 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a0ddaae-cb65-45e0-a82a-44f6afb10d48" path="/var/lib/kubelet/pods/8a0ddaae-cb65-45e0-a82a-44f6afb10d48/volumes" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.708488 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a90441f8-30d7-4091-a23f-2522c586f8cf" path="/var/lib/kubelet/pods/a90441f8-30d7-4091-a23f-2522c586f8cf/volumes" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.708890 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7ffd372-5054-48ec-8195-dc89cf096f38" path="/var/lib/kubelet/pods/f7ffd372-5054-48ec-8195-dc89cf096f38/volumes" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.710831 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.711121 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.711399 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-catalog-content\") pod \"community-operators-rq699\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " pod="openshift-marketplace/community-operators-rq699" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.711427 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-utilities\") pod \"community-operators-rq699\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " pod="openshift-marketplace/community-operators-rq699" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.711451 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nm8p\" (UniqueName: \"kubernetes.io/projected/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-kube-api-access-6nm8p\") pod \"community-operators-rq699\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " pod="openshift-marketplace/community-operators-rq699" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.712021 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-catalog-content\") pod \"community-operators-rq699\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " pod="openshift-marketplace/community-operators-rq699" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.712077 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-utilities\") pod \"community-operators-rq699\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " pod="openshift-marketplace/community-operators-rq699" Mar 17 15:40:21 crc kubenswrapper[4767]: E0317 15:40:21.713048 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:22.213028994 +0000 UTC m=+213.626345031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.715204 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8lw6\" (UniqueName: \"kubernetes.io/projected/3dc430ad-a22a-4fe8-a4dd-5c94622da680-kube-api-access-r8lw6\") pod \"community-operators-klkzj\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.739384 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.747284 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nm8p\" (UniqueName: \"kubernetes.io/projected/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-kube-api-access-6nm8p\") pod \"community-operators-rq699\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " pod="openshift-marketplace/community-operators-rq699" Mar 17 15:40:21 crc kubenswrapper[4767]: I0317 15:40:21.748051 4767 ???:1] "http: TLS handshake error from 192.168.126.11:35812: no serving certificate available for the kubelet" Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.084896 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rq699" Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.099456 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:22 crc kubenswrapper[4767]: E0317 15:40:22.099925 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:22.599906551 +0000 UTC m=+214.013222598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.144480 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-frl2d" event={"ID":"201932aa-c539-434b-992e-92f889e52de3","Type":"ContainerStarted","Data":"0687c4ccf67e9662c1c2f3adfabb002b4e9991632e3a881741db5c7e41463f52"} Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.144829 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-676bc56c59-88gvd" Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.145334 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.145392 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.165829 4767 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.198667 4767 ???:1] "http: TLS handshake error from 192.168.126.11:35818: no serving certificate available for the kubelet" Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.204037 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:22 crc kubenswrapper[4767]: E0317 15:40:22.204272 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:22.704251113 +0000 UTC m=+214.117567160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.204494 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:22 crc kubenswrapper[4767]: E0317 15:40:22.205235 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:22.705222937 +0000 UTC m=+214.118538984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.280743 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:22 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:22 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:22 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.280804 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.789913 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:22 crc kubenswrapper[4767]: E0317 15:40:22.790387 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:23.790367539 +0000 UTC m=+215.203683606 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.898479 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:22 crc kubenswrapper[4767]: E0317 15:40:22.901776 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:23.40175061 +0000 UTC m=+214.815066657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.926497 4767 ???:1] "http: TLS handshake error from 192.168.126.11:35822: no serving certificate available for the kubelet" Mar 17 15:40:22 crc kubenswrapper[4767]: I0317 15:40:22.968370 4767 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 17 15:40:23 crc kubenswrapper[4767]: I0317 15:40:23.118448 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:23 crc kubenswrapper[4767]: E0317 15:40:23.126495 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 15:40:23.62647119 +0000 UTC m=+215.039787237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ldbnt" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:23 crc kubenswrapper[4767]: I0317 15:40:23.172586 4767 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-17T15:40:22.968399888Z","Handler":null,"Name":""} Mar 17 15:40:23 crc kubenswrapper[4767]: I0317 15:40:23.639348 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:23 crc kubenswrapper[4767]: E0317 15:40:23.640024 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 15:40:24.640007159 +0000 UTC m=+216.053323196 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 15:40:23 crc kubenswrapper[4767]: I0317 15:40:23.641248 4767 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 17 15:40:23 crc kubenswrapper[4767]: I0317 15:40:23.641275 4767 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 17 15:40:23 crc kubenswrapper[4767]: I0317 15:40:23.647862 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c7ktd"] Mar 17 15:40:23 crc kubenswrapper[4767]: I0317 15:40:23.664646 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:40:23 crc kubenswrapper[4767]: I0317 15:40:23.929787 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 15:40:23 crc kubenswrapper[4767]: I0317 15:40:23.961468 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:23 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:23 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:23 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.143807 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.155719 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.182704 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.182735 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zxsb9"] Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.183841 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q2n5r"] Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.183931 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.184841 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7ktd"] Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.184900 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.190309 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-utilities\") pod \"redhat-marketplace-c7ktd\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.190383 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2dfp\" (UniqueName: \"kubernetes.io/projected/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-kube-api-access-q2dfp\") pod \"redhat-marketplace-c7ktd\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.190421 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-catalog-content\") pod \"redhat-marketplace-c7ktd\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.190583 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.211965 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxsb9"] Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.222216 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.250367 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q2n5r"] Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.267197 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vhw26"] Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.273822 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.291887 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-utilities\") pod \"redhat-marketplace-zxsb9\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.291945 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdbm5\" (UniqueName: \"kubernetes.io/projected/ebb17b29-f33c-4236-bb9d-034821b17ba3-kube-api-access-cdbm5\") pod \"redhat-operators-vhw26\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.291995 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-catalog-content\") pod \"redhat-operators-q2n5r\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.292077 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-utilities\") pod \"redhat-marketplace-c7ktd\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.292103 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-catalog-content\") pod \"redhat-marketplace-zxsb9\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.292134 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.296819 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2dfp\" (UniqueName: \"kubernetes.io/projected/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-kube-api-access-q2dfp\") pod \"redhat-marketplace-c7ktd\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.296929 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-catalog-content\") pod \"redhat-marketplace-c7ktd\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.296975 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc7fv\" (UniqueName: \"kubernetes.io/projected/e76337aa-429a-49b9-99d3-e261c32ee85b-kube-api-access-xc7fv\") pod \"redhat-operators-q2n5r\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.297085 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-utilities\") pod \"redhat-operators-vhw26\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.297147 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhfdf\" (UniqueName: \"kubernetes.io/projected/3f4f4f7f-0983-4551-9a1c-88c6214322f4-kube-api-access-xhfdf\") pod \"redhat-marketplace-zxsb9\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.297248 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-utilities\") pod \"redhat-operators-q2n5r\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.297293 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-catalog-content\") pod \"redhat-operators-vhw26\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.299601 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vhw26"] Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.301703 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-utilities\") pod \"redhat-marketplace-c7ktd\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.307689 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-catalog-content\") pod \"redhat-marketplace-c7ktd\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.316624 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.316671 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.355893 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2dfp\" (UniqueName: \"kubernetes.io/projected/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-kube-api-access-q2dfp\") pod \"redhat-marketplace-c7ktd\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.391923 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:24 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:24 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:24 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.392299 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.398985 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-utilities\") pod \"redhat-operators-q2n5r\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.399021 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-catalog-content\") pod \"redhat-operators-vhw26\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.399059 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-utilities\") pod \"redhat-marketplace-zxsb9\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.399083 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdbm5\" (UniqueName: \"kubernetes.io/projected/ebb17b29-f33c-4236-bb9d-034821b17ba3-kube-api-access-cdbm5\") pod \"redhat-operators-vhw26\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.399113 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-catalog-content\") pod \"redhat-operators-q2n5r\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.399138 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-catalog-content\") pod \"redhat-marketplace-zxsb9\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.399209 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc7fv\" (UniqueName: \"kubernetes.io/projected/e76337aa-429a-49b9-99d3-e261c32ee85b-kube-api-access-xc7fv\") pod \"redhat-operators-q2n5r\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.399246 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-utilities\") pod \"redhat-operators-vhw26\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.399269 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhfdf\" (UniqueName: \"kubernetes.io/projected/3f4f4f7f-0983-4551-9a1c-88c6214322f4-kube-api-access-xhfdf\") pod \"redhat-marketplace-zxsb9\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.400933 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-catalog-content\") pod \"redhat-marketplace-zxsb9\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.402537 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-utilities\") pod \"redhat-operators-q2n5r\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.403029 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-catalog-content\") pod \"redhat-operators-vhw26\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.403777 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-utilities\") pod \"redhat-marketplace-zxsb9\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.404835 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-utilities\") pod \"redhat-operators-vhw26\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.405295 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-catalog-content\") pod \"redhat-operators-q2n5r\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.538610 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:40:24 crc kubenswrapper[4767]: E0317 15:40:24.636441 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-podba34e593_ec5d_4e7e_9505_e21f3242f3b7.slice/crio-853074c8dd907aad73979fb890b19677646d3dd46327a4c6c7bc585d1b71bd09.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-podba34e593_ec5d_4e7e_9505_e21f3242f3b7.slice/crio-conmon-853074c8dd907aad73979fb890b19677646d3dd46327a4c6c7bc585d1b71bd09.scope\": RecentStats: unable to find data in memory cache]" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.839645 4767 ???:1] "http: TLS handshake error from 192.168.126.11:35836: no serving certificate available for the kubelet" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.851575 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc7fv\" (UniqueName: \"kubernetes.io/projected/e76337aa-429a-49b9-99d3-e261c32ee85b-kube-api-access-xc7fv\") pod \"redhat-operators-q2n5r\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.859216 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ldbnt\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.867656 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhfdf\" (UniqueName: \"kubernetes.io/projected/3f4f4f7f-0983-4551-9a1c-88c6214322f4-kube-api-access-xhfdf\") pod \"redhat-marketplace-zxsb9\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.874013 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdbm5\" (UniqueName: \"kubernetes.io/projected/ebb17b29-f33c-4236-bb9d-034821b17ba3-kube-api-access-cdbm5\") pod \"redhat-operators-vhw26\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.884225 4767 generic.go:334] "Generic (PLEG): container finished" podID="ba34e593-ec5d-4e7e-9505-e21f3242f3b7" containerID="853074c8dd907aad73979fb890b19677646d3dd46327a4c6c7bc585d1b71bd09" exitCode=0 Mar 17 15:40:24 crc kubenswrapper[4767]: I0317 15:40:24.884295 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ba34e593-ec5d-4e7e-9505-e21f3242f3b7","Type":"ContainerDied","Data":"853074c8dd907aad73979fb890b19677646d3dd46327a4c6c7bc585d1b71bd09"} Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.053565 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.059610 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.071984 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-klkzj"] Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.092474 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.146289 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.163086 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d2kvv"] Mar 17 15:40:25 crc kubenswrapper[4767]: W0317 15:40:25.191179 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ee8aaaa_9a6c_4ef6_9adb_b8f0ee50ee9a.slice/crio-1c5cba4c2d6d7f1607b59467bc764d17bc2be44476f3445ee25fec9c51f99e3d WatchSource:0}: Error finding container 1c5cba4c2d6d7f1607b59467bc764d17bc2be44476f3445ee25fec9c51f99e3d: Status 404 returned error can't find the container with id 1c5cba4c2d6d7f1607b59467bc764d17bc2be44476f3445ee25fec9c51f99e3d Mar 17 15:40:25 crc kubenswrapper[4767]: W0317 15:40:25.191812 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dc430ad_a22a_4fe8_a4dd_5c94622da680.slice/crio-335a92007e3345a9487740053d8b2f9128b93449780abe80f96ae75127834f06 WatchSource:0}: Error finding container 335a92007e3345a9487740053d8b2f9128b93449780abe80f96ae75127834f06: Status 404 returned error can't find the container with id 335a92007e3345a9487740053d8b2f9128b93449780abe80f96ae75127834f06 Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.205777 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.267741 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:25 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:25 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:25 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.267807 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.297033 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rq699"] Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.324395 4767 patch_prober.go:28] interesting pod/apiserver-76f77b778f-bk7p5 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]log ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]etcd ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/generic-apiserver-start-informers ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/max-in-flight-filter ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 17 15:40:25 crc kubenswrapper[4767]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 17 15:40:25 crc kubenswrapper[4767]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/project.openshift.io-projectcache ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/openshift.io-startinformers ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 17 15:40:25 crc kubenswrapper[4767]: livez check failed Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.324625 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" podUID="1c88a6ee-cd24-4d85-8a89-d830e5baa434" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.434951 4767 patch_prober.go:28] interesting pod/apiserver-76f77b778f-bk7p5 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]log ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]etcd ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/generic-apiserver-start-informers ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/max-in-flight-filter ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 17 15:40:25 crc kubenswrapper[4767]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 17 15:40:25 crc kubenswrapper[4767]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/project.openshift.io-projectcache ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/openshift.io-startinformers ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 17 15:40:25 crc kubenswrapper[4767]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 17 15:40:25 crc kubenswrapper[4767]: livez check failed Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.435029 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" podUID="1c88a6ee-cd24-4d85-8a89-d830e5baa434" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.461516 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.623544 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66dc66f96c-x2grn"] Mar 17 15:40:25 crc kubenswrapper[4767]: I0317 15:40:25.922930 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7ktd"] Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:25.929112 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-79fk6"] Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.102488 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"97a26755-5343-4d6b-92b1-a05b5542944f","Type":"ContainerDied","Data":"faa4b40271082a42499613a1f75213c9735de220ba7b688fa9d162245af63f6c"} Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.131298 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs"] Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.102440 4767 generic.go:334] "Generic (PLEG): container finished" podID="97a26755-5343-4d6b-92b1-a05b5542944f" containerID="faa4b40271082a42499613a1f75213c9735de220ba7b688fa9d162245af63f6c" exitCode=0 Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.179702 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2kvv" event={"ID":"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a","Type":"ContainerStarted","Data":"1c5cba4c2d6d7f1607b59467bc764d17bc2be44476f3445ee25fec9c51f99e3d"} Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.195345 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" event={"ID":"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1","Type":"ContainerStarted","Data":"8fe19b3bcd146dbf471e425a36f069ddb1b2b609b9cc9fbebdf009b6dee4a058"} Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.233571 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq699" event={"ID":"7d1d71c5-161c-4ba3-96da-ed07128ff6e1","Type":"ContainerStarted","Data":"5eb196e2a1995f63348cd0dbf9c014ae3277f9a4170460732e92d81f4382a2e2"} Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.272935 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-frl2d" event={"ID":"201932aa-c539-434b-992e-92f889e52de3","Type":"ContainerStarted","Data":"1c7eafa4bce86e0d31f4125507e2400c0b7865303d2efc7b3d0f115c7f2d4771"} Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.275783 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:26 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:26 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:26 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.275816 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.277504 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkzj" event={"ID":"3dc430ad-a22a-4fe8-a4dd-5c94622da680","Type":"ContainerStarted","Data":"335a92007e3345a9487740053d8b2f9128b93449780abe80f96ae75127834f06"} Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.279689 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q2n5r"] Mar 17 15:40:26 crc kubenswrapper[4767]: W0317 15:40:26.313466 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode76337aa_429a_49b9_99d3_e261c32ee85b.slice/crio-debde7dee9f9c11a29fffe2fda7c1ab4623535bd1b88b83d2e6f1407cec24113 WatchSource:0}: Error finding container debde7dee9f9c11a29fffe2fda7c1ab4623535bd1b88b83d2e6f1407cec24113: Status 404 returned error can't find the container with id debde7dee9f9c11a29fffe2fda7c1ab4623535bd1b88b83d2e6f1407cec24113 Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.318927 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-frl2d" podStartSLOduration=40.318907425 podStartE2EDuration="40.318907425s" podCreationTimestamp="2026-03-17 15:39:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:26.309042066 +0000 UTC m=+217.722358123" watchObservedRunningTime="2026-03-17 15:40:26.318907425 +0000 UTC m=+217.732223472" Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.778432 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxsb9"] Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.835822 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vhw26"] Mar 17 15:40:26 crc kubenswrapper[4767]: I0317 15:40:26.886774 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ldbnt"] Mar 17 15:40:26 crc kubenswrapper[4767]: W0317 15:40:26.921146 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb17b29_f33c_4236_bb9d_034821b17ba3.slice/crio-181e06f5e1d81e1cc3b53ceb2a2dec30da8fe62b7322f74b32a24c63d584d7cb WatchSource:0}: Error finding container 181e06f5e1d81e1cc3b53ceb2a2dec30da8fe62b7322f74b32a24c63d584d7cb: Status 404 returned error can't find the container with id 181e06f5e1d81e1cc3b53ceb2a2dec30da8fe62b7322f74b32a24c63d584d7cb Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.306764 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" event={"ID":"83518d71-dc9f-4e6c-9ad4-66ee45c262f8","Type":"ContainerStarted","Data":"74361c254ab87b351356f4fee0255dc0c45e89d710e2fdddefc488a754694d35"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.307945 4767 generic.go:334] "Generic (PLEG): container finished" podID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerID="1856ee116c91f280b647868400d6c8522c8047bf2bc8ac741f920258bc1e964d" exitCode=0 Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.307982 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq699" event={"ID":"7d1d71c5-161c-4ba3-96da-ed07128ff6e1","Type":"ContainerDied","Data":"1856ee116c91f280b647868400d6c8522c8047bf2bc8ac741f920258bc1e964d"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.311844 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" event={"ID":"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1","Type":"ContainerStarted","Data":"c2ff9cf9ceadefea20ba0614ca9aa92e18891d8de41b3626a512d5f95a723dd1"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.312808 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.314291 4767 patch_prober.go:28] interesting pod/controller-manager-66dc66f96c-x2grn container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" start-of-body= Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.314319 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" podUID="f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.329123 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:27 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:27 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:27 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.329206 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.329937 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2n5r" event={"ID":"e76337aa-429a-49b9-99d3-e261c32ee85b","Type":"ContainerStarted","Data":"debde7dee9f9c11a29fffe2fda7c1ab4623535bd1b88b83d2e6f1407cec24113"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.338445 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhw26" event={"ID":"ebb17b29-f33c-4236-bb9d-034821b17ba3","Type":"ContainerStarted","Data":"181e06f5e1d81e1cc3b53ceb2a2dec30da8fe62b7322f74b32a24c63d584d7cb"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.341141 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79fk6" event={"ID":"5df92c46-85d1-4900-9916-486aa896d42a","Type":"ContainerStarted","Data":"fb70c784f417fb6c4726bfcd62f807ec81c6f6f9300e3ea14facd1c59d2bdb62"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.352309 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ba34e593-ec5d-4e7e-9505-e21f3242f3b7","Type":"ContainerDied","Data":"d6484187f76ec5fb1c4fde08f54929198dac176e5ebb78078768ffda550573ca"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.352354 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6484187f76ec5fb1c4fde08f54929198dac176e5ebb78078768ffda550573ca" Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.359766 4767 generic.go:334] "Generic (PLEG): container finished" podID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" containerID="ac6c43a4f582c64a7a75cd098f8481ec8534a13bc07188100494824bd66afd48" exitCode=0 Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.360120 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.362420 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkzj" event={"ID":"3dc430ad-a22a-4fe8-a4dd-5c94622da680","Type":"ContainerDied","Data":"ac6c43a4f582c64a7a75cd098f8481ec8534a13bc07188100494824bd66afd48"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.365838 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxsb9" event={"ID":"3f4f4f7f-0983-4551-9a1c-88c6214322f4","Type":"ContainerStarted","Data":"b01bb10f584c109c735f4dc8778b8716940f893878b44dfd96eabee8babfdb90"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.370131 4767 generic.go:334] "Generic (PLEG): container finished" podID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" containerID="3a8ad9a791ce5935f91019adfaf3fbd5123b1ffc65d7e28e0021f28c1ed1b7f5" exitCode=0 Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.370601 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2kvv" event={"ID":"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a","Type":"ContainerDied","Data":"3a8ad9a791ce5935f91019adfaf3fbd5123b1ffc65d7e28e0021f28c1ed1b7f5"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.373681 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" event={"ID":"c3569d92-3c22-4a58-b535-0400b94005a2","Type":"ContainerStarted","Data":"7b41dc2703bda2c0a1f1650e31665a3976cf998ee08246863d3c310aeeadf46f"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.383508 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7ktd" event={"ID":"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3","Type":"ContainerStarted","Data":"92057d188bd11eb864de3b3a7e87ef9b1e42b9728ae958e64b6665ccb4dda19c"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.383560 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7ktd" event={"ID":"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3","Type":"ContainerStarted","Data":"fa8520c4808219b785e6d448797823dd736e25bc5358b98a2840588a5ee52388"} Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.384552 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" podStartSLOduration=7.384536486 podStartE2EDuration="7.384536486s" podCreationTimestamp="2026-03-17 15:40:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:27.381456997 +0000 UTC m=+218.794773064" watchObservedRunningTime="2026-03-17 15:40:27.384536486 +0000 UTC m=+218.797852533" Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.528002 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kubelet-dir\") pod \"ba34e593-ec5d-4e7e-9505-e21f3242f3b7\" (UID: \"ba34e593-ec5d-4e7e-9505-e21f3242f3b7\") " Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.528094 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ba34e593-ec5d-4e7e-9505-e21f3242f3b7" (UID: "ba34e593-ec5d-4e7e-9505-e21f3242f3b7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.529412 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kube-api-access\") pod \"ba34e593-ec5d-4e7e-9505-e21f3242f3b7\" (UID: \"ba34e593-ec5d-4e7e-9505-e21f3242f3b7\") " Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.534709 4767 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.541942 4767 ???:1] "http: TLS handshake error from 192.168.126.11:35842: no serving certificate available for the kubelet" Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.552746 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ba34e593-ec5d-4e7e-9505-e21f3242f3b7" (UID: "ba34e593-ec5d-4e7e-9505-e21f3242f3b7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:40:27 crc kubenswrapper[4767]: I0317 15:40:27.638551 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ba34e593-ec5d-4e7e-9505-e21f3242f3b7-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.106704 4767 patch_prober.go:28] interesting pod/console-f9d7485db-9cz8q container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.106760 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9cz8q" podUID="deba5665-b026-4d87-a986-b17f02f644d7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.268421 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:28 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:28 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:28 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.269279 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.270984 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.271014 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.273189 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.273242 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.629158 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" event={"ID":"83518d71-dc9f-4e6c-9ad4-66ee45c262f8","Type":"ContainerStarted","Data":"3e2487e816fde5f768d7d7f328bda10584be833b3a25b5b7497b053e59a49751"} Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.629613 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.719307 4767 generic.go:334] "Generic (PLEG): container finished" podID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerID="92057d188bd11eb864de3b3a7e87ef9b1e42b9728ae958e64b6665ccb4dda19c" exitCode=0 Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.719417 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7ktd" event={"ID":"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3","Type":"ContainerDied","Data":"92057d188bd11eb864de3b3a7e87ef9b1e42b9728ae958e64b6665ccb4dda19c"} Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.727490 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"97a26755-5343-4d6b-92b1-a05b5542944f","Type":"ContainerDied","Data":"6716b310725f103621b887bacafbf1e8cbaa7d078d4bde96b5561d70c270499a"} Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.738112 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6716b310725f103621b887bacafbf1e8cbaa7d078d4bde96b5561d70c270499a" Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.738256 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxsb9" event={"ID":"3f4f4f7f-0983-4551-9a1c-88c6214322f4","Type":"ContainerDied","Data":"15c6634662eb24ce7a354c47d11199fe27a893e241b63dc425e5977b70a2a0ed"} Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.732041 4767 generic.go:334] "Generic (PLEG): container finished" podID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" containerID="15c6634662eb24ce7a354c47d11199fe27a893e241b63dc425e5977b70a2a0ed" exitCode=0 Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.758765 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" podStartSLOduration=154.758750574 podStartE2EDuration="2m34.758750574s" podCreationTimestamp="2026-03-17 15:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:28.756546986 +0000 UTC m=+220.169863053" watchObservedRunningTime="2026-03-17 15:40:28.758750574 +0000 UTC m=+220.172066621" Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.762141 4767 generic.go:334] "Generic (PLEG): container finished" podID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerID="c75e05a67860ba455fba7d2f51422da56388bdc9c2e87ce454f03509b35ce552" exitCode=0 Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.763227 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2n5r" event={"ID":"e76337aa-429a-49b9-99d3-e261c32ee85b","Type":"ContainerDied","Data":"c75e05a67860ba455fba7d2f51422da56388bdc9c2e87ce454f03509b35ce552"} Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.775871 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" event={"ID":"c3569d92-3c22-4a58-b535-0400b94005a2","Type":"ContainerStarted","Data":"6e41f5befd636324890be40e1c382e04ad5a319fa271921f6130bd97e3b10b38"} Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.779631 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.783419 4767 generic.go:334] "Generic (PLEG): container finished" podID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerID="e3820da4606dbc685eb002165ce2ce61f0b8d5cde93ec9d4290826cde4f4e9fb" exitCode=0 Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.783637 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhw26" event={"ID":"ebb17b29-f33c-4236-bb9d-034821b17ba3","Type":"ContainerDied","Data":"e3820da4606dbc685eb002165ce2ce61f0b8d5cde93ec9d4290826cde4f4e9fb"} Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.791624 4767 generic.go:334] "Generic (PLEG): container finished" podID="5df92c46-85d1-4900-9916-486aa896d42a" containerID="ff528ef531b732270c2a7159eabb44c8387e17ced0fcef2301dd1798a7cac3d3" exitCode=0 Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.791764 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 15:40:28 crc kubenswrapper[4767]: I0317 15:40:28.823273 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79fk6" event={"ID":"5df92c46-85d1-4900-9916-486aa896d42a","Type":"ContainerDied","Data":"ff528ef531b732270c2a7159eabb44c8387e17ced0fcef2301dd1798a7cac3d3"} Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.135512 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.161042 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.237392 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97a26755-5343-4d6b-92b1-a05b5542944f-kubelet-dir\") pod \"97a26755-5343-4d6b-92b1-a05b5542944f\" (UID: \"97a26755-5343-4d6b-92b1-a05b5542944f\") " Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.237480 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97a26755-5343-4d6b-92b1-a05b5542944f-kube-api-access\") pod \"97a26755-5343-4d6b-92b1-a05b5542944f\" (UID: \"97a26755-5343-4d6b-92b1-a05b5542944f\") " Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.237521 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97a26755-5343-4d6b-92b1-a05b5542944f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "97a26755-5343-4d6b-92b1-a05b5542944f" (UID: "97a26755-5343-4d6b-92b1-a05b5542944f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.238021 4767 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97a26755-5343-4d6b-92b1-a05b5542944f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.241073 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" podStartSLOduration=9.241045287 podStartE2EDuration="9.241045287s" podCreationTimestamp="2026-03-17 15:40:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:29.229774478 +0000 UTC m=+220.643090545" watchObservedRunningTime="2026-03-17 15:40:29.241045287 +0000 UTC m=+220.654361334" Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.270319 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:29 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:29 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:29 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.270386 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.355701 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97a26755-5343-4d6b-92b1-a05b5542944f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "97a26755-5343-4d6b-92b1-a05b5542944f" (UID: "97a26755-5343-4d6b-92b1-a05b5542944f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.418890 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.446376 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97a26755-5343-4d6b-92b1-a05b5542944f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.626513 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:40:29 crc kubenswrapper[4767]: I0317 15:40:29.646874 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" Mar 17 15:40:30 crc kubenswrapper[4767]: I0317 15:40:30.012319 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 15:40:30 crc kubenswrapper[4767]: I0317 15:40:30.410695 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:30 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:30 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:30 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:30 crc kubenswrapper[4767]: I0317 15:40:30.410748 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:31 crc kubenswrapper[4767]: I0317 15:40:31.273381 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:31 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:31 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:31 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:31 crc kubenswrapper[4767]: I0317 15:40:31.273471 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:32 crc kubenswrapper[4767]: I0317 15:40:32.779003 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:40:32 crc kubenswrapper[4767]: I0317 15:40:32.779908 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:40:32 crc kubenswrapper[4767]: I0317 15:40:32.780037 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:40:32 crc kubenswrapper[4767]: I0317 15:40:32.780080 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:40:32 crc kubenswrapper[4767]: I0317 15:40:32.786167 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 17 15:40:32 crc kubenswrapper[4767]: I0317 15:40:32.786489 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 17 15:40:32 crc kubenswrapper[4767]: I0317 15:40:32.786581 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:32 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:32 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:32 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:32 crc kubenswrapper[4767]: I0317 15:40:32.786636 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:32 crc kubenswrapper[4767]: I0317 15:40:32.796802 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:40:32 crc kubenswrapper[4767]: I0317 15:40:32.849985 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 17 15:40:32 crc kubenswrapper[4767]: I0317 15:40:32.851044 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 17 15:40:33 crc kubenswrapper[4767]: I0317 15:40:33.000585 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:40:33 crc kubenswrapper[4767]: I0317 15:40:33.001469 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:40:33 crc kubenswrapper[4767]: I0317 15:40:33.019857 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:40:33 crc kubenswrapper[4767]: I0317 15:40:33.026882 4767 ???:1] "http: TLS handshake error from 192.168.126.11:47048: no serving certificate available for the kubelet" Mar 17 15:40:33 crc kubenswrapper[4767]: I0317 15:40:33.056814 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 15:40:33 crc kubenswrapper[4767]: I0317 15:40:33.294960 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:40:33 crc kubenswrapper[4767]: I0317 15:40:33.295787 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 15:40:33 crc kubenswrapper[4767]: I0317 15:40:33.304450 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:33 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:33 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:33 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:33 crc kubenswrapper[4767]: I0317 15:40:33.304526 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:34 crc kubenswrapper[4767]: I0317 15:40:34.448412 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:34 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:34 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:34 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:34 crc kubenswrapper[4767]: I0317 15:40:34.448793 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:34 crc kubenswrapper[4767]: I0317 15:40:34.501375 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:40:34 crc kubenswrapper[4767]: I0317 15:40:34.501461 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:40:35 crc kubenswrapper[4767]: I0317 15:40:35.576448 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:35 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:35 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:35 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:35 crc kubenswrapper[4767]: I0317 15:40:35.576878 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:36 crc kubenswrapper[4767]: I0317 15:40:36.385405 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 15:40:36 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Mar 17 15:40:36 crc kubenswrapper[4767]: [+]process-running ok Mar 17 15:40:36 crc kubenswrapper[4767]: healthz check failed Mar 17 15:40:36 crc kubenswrapper[4767]: I0317 15:40:36.385458 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 15:40:37 crc kubenswrapper[4767]: I0317 15:40:37.754341 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:40:37 crc kubenswrapper[4767]: I0317 15:40:37.758725 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 15:40:37 crc kubenswrapper[4767]: I0317 15:40:37.879666 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"52868f140393f4f82c728c14f88547fa8c5a8e9cc9a1400b09a2d660651ea126"} Mar 17 15:40:37 crc kubenswrapper[4767]: I0317 15:40:37.894330 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a5116a97d5cb5b0074e771f8bf326621035d709685d80182d4ba7ff0c7a516f0"} Mar 17 15:40:37 crc kubenswrapper[4767]: I0317 15:40:37.897293 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"07a96146450f531ff7786b2da8beb44a3b2606f34594641d3972da59b20fe2a9"} Mar 17 15:40:38 crc kubenswrapper[4767]: I0317 15:40:38.052945 4767 patch_prober.go:28] interesting pod/console-f9d7485db-9cz8q container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 17 15:40:38 crc kubenswrapper[4767]: I0317 15:40:38.053004 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9cz8q" podUID="deba5665-b026-4d87-a986-b17f02f644d7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 17 15:40:38 crc kubenswrapper[4767]: I0317 15:40:38.289227 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:38 crc kubenswrapper[4767]: I0317 15:40:38.289552 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:38 crc kubenswrapper[4767]: I0317 15:40:38.289431 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:38 crc kubenswrapper[4767]: I0317 15:40:38.289624 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:38 crc kubenswrapper[4767]: I0317 15:40:38.993325 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"668abd9c68f53f387f84ed857c59b54d536b82372787e441c943d13d8627e07c"} Mar 17 15:40:39 crc kubenswrapper[4767]: I0317 15:40:39.115512 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6cd54d87340754c93f2d193b978b4567b4162245c54f66457183c97595c6a048"} Mar 17 15:40:39 crc kubenswrapper[4767]: I0317 15:40:39.115682 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:40:39 crc kubenswrapper[4767]: I0317 15:40:39.243988 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"27ec17e2bf9c98886e7dcf736db7622ccd7f468a4577e4f71a6efda4c5065fbf"} Mar 17 15:40:39 crc kubenswrapper[4767]: I0317 15:40:39.583834 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.151682 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 17 15:40:41 crc kubenswrapper[4767]: E0317 15:40:41.151935 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba34e593-ec5d-4e7e-9505-e21f3242f3b7" containerName="pruner" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.151947 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba34e593-ec5d-4e7e-9505-e21f3242f3b7" containerName="pruner" Mar 17 15:40:41 crc kubenswrapper[4767]: E0317 15:40:41.151957 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a26755-5343-4d6b-92b1-a05b5542944f" containerName="pruner" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.151965 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a26755-5343-4d6b-92b1-a05b5542944f" containerName="pruner" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.152075 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba34e593-ec5d-4e7e-9505-e21f3242f3b7" containerName="pruner" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.152085 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="97a26755-5343-4d6b-92b1-a05b5542944f" containerName="pruner" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.152740 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.156559 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.156669 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.171451 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.693035 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/177d0275-072c-4e59-8c1a-4017361bf42f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"177d0275-072c-4e59-8c1a-4017361bf42f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.693982 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/177d0275-072c-4e59-8c1a-4017361bf42f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"177d0275-072c-4e59-8c1a-4017361bf42f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.799492 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/177d0275-072c-4e59-8c1a-4017361bf42f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"177d0275-072c-4e59-8c1a-4017361bf42f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.799619 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/177d0275-072c-4e59-8c1a-4017361bf42f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"177d0275-072c-4e59-8c1a-4017361bf42f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.799769 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/177d0275-072c-4e59-8c1a-4017361bf42f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"177d0275-072c-4e59-8c1a-4017361bf42f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 15:40:41 crc kubenswrapper[4767]: I0317 15:40:41.844407 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/177d0275-072c-4e59-8c1a-4017361bf42f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"177d0275-072c-4e59-8c1a-4017361bf42f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 15:40:42 crc kubenswrapper[4767]: I0317 15:40:42.133337 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 15:40:43 crc kubenswrapper[4767]: I0317 15:40:43.340047 4767 ???:1] "http: TLS handshake error from 192.168.126.11:39612: no serving certificate available for the kubelet" Mar 17 15:40:43 crc kubenswrapper[4767]: I0317 15:40:43.402098 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 17 15:40:43 crc kubenswrapper[4767]: I0317 15:40:43.790285 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"177d0275-072c-4e59-8c1a-4017361bf42f","Type":"ContainerStarted","Data":"7c8f0ae6ddd5b6baca5c5c17c2c0e3118ca1dfd361b00d432293a618d6362d9a"} Mar 17 15:40:44 crc kubenswrapper[4767]: I0317 15:40:44.880160 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"177d0275-072c-4e59-8c1a-4017361bf42f","Type":"ContainerStarted","Data":"9219b42482feef2e2e0a1102dafb633834792442022f79c8b446986ac59b9300"} Mar 17 15:40:44 crc kubenswrapper[4767]: I0317 15:40:44.908652 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=3.908631631 podStartE2EDuration="3.908631631s" podCreationTimestamp="2026-03-17 15:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:40:44.908143104 +0000 UTC m=+236.321459161" watchObservedRunningTime="2026-03-17 15:40:44.908631631 +0000 UTC m=+236.321947688" Mar 17 15:40:45 crc kubenswrapper[4767]: I0317 15:40:45.102689 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:40:46 crc kubenswrapper[4767]: I0317 15:40:46.977541 4767 generic.go:334] "Generic (PLEG): container finished" podID="177d0275-072c-4e59-8c1a-4017361bf42f" containerID="9219b42482feef2e2e0a1102dafb633834792442022f79c8b446986ac59b9300" exitCode=0 Mar 17 15:40:46 crc kubenswrapper[4767]: I0317 15:40:46.977644 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"177d0275-072c-4e59-8c1a-4017361bf42f","Type":"ContainerDied","Data":"9219b42482feef2e2e0a1102dafb633834792442022f79c8b446986ac59b9300"} Mar 17 15:40:47 crc kubenswrapper[4767]: I0317 15:40:47.434129 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 17 15:40:47 crc kubenswrapper[4767]: I0317 15:40:47.435577 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:40:47 crc kubenswrapper[4767]: I0317 15:40:47.451738 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 17 15:40:47 crc kubenswrapper[4767]: I0317 15:40:47.662817 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:40:47 crc kubenswrapper[4767]: I0317 15:40:47.662889 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-var-lock\") pod \"installer-9-crc\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:40:47 crc kubenswrapper[4767]: I0317 15:40:47.662931 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d49401a4-1f38-4e26-92d4-69c2085d6640-kube-api-access\") pod \"installer-9-crc\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:40:47 crc kubenswrapper[4767]: I0317 15:40:47.765142 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:40:47 crc kubenswrapper[4767]: I0317 15:40:47.765236 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-var-lock\") pod \"installer-9-crc\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:40:47 crc kubenswrapper[4767]: I0317 15:40:47.765275 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d49401a4-1f38-4e26-92d4-69c2085d6640-kube-api-access\") pod \"installer-9-crc\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:40:47 crc kubenswrapper[4767]: I0317 15:40:47.765651 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:40:47 crc kubenswrapper[4767]: I0317 15:40:47.765909 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-var-lock\") pod \"installer-9-crc\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.086883 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d49401a4-1f38-4e26-92d4-69c2085d6640-kube-api-access\") pod \"installer-9-crc\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.090312 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.105687 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.271874 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.272364 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.272507 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-59hsl" Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.272202 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.273694 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.273980 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.274023 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.274536 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"140aaa6eab7b06985ec3865ee30e1ab9ae5685144782d645d3c774697c868c33"} pod="openshift-console/downloads-7954f5f757-59hsl" containerMessage="Container download-server failed liveness probe, will be restarted" Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.274602 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" containerID="cri-o://140aaa6eab7b06985ec3865ee30e1ab9ae5685144782d645d3c774697c868c33" gracePeriod=2 Mar 17 15:40:48 crc kubenswrapper[4767]: I0317 15:40:48.372486 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:40:49 crc kubenswrapper[4767]: I0317 15:40:49.131293 4767 generic.go:334] "Generic (PLEG): container finished" podID="471b6054-3b2a-47c9-8889-942603600da9" containerID="140aaa6eab7b06985ec3865ee30e1ab9ae5685144782d645d3c774697c868c33" exitCode=0 Mar 17 15:40:49 crc kubenswrapper[4767]: I0317 15:40:49.131647 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-59hsl" event={"ID":"471b6054-3b2a-47c9-8889-942603600da9","Type":"ContainerDied","Data":"140aaa6eab7b06985ec3865ee30e1ab9ae5685144782d645d3c774697c868c33"} Mar 17 15:40:49 crc kubenswrapper[4767]: I0317 15:40:49.131735 4767 scope.go:117] "RemoveContainer" containerID="c82a98015caa0bb7c218c4c3adb06778f84051c358433937ea7560b971a25552" Mar 17 15:40:58 crc kubenswrapper[4767]: I0317 15:40:58.274649 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:40:58 crc kubenswrapper[4767]: I0317 15:40:58.275851 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:41:03 crc kubenswrapper[4767]: I0317 15:41:03.857751 4767 ???:1] "http: TLS handshake error from 192.168.126.11:42000: no serving certificate available for the kubelet" Mar 17 15:41:04 crc kubenswrapper[4767]: I0317 15:41:04.172324 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:41:04 crc kubenswrapper[4767]: I0317 15:41:04.172383 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:41:08 crc kubenswrapper[4767]: I0317 15:41:08.271351 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:41:08 crc kubenswrapper[4767]: I0317 15:41:08.271696 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:41:10 crc kubenswrapper[4767]: I0317 15:41:10.124539 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 15:41:10 crc kubenswrapper[4767]: I0317 15:41:10.232456 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/177d0275-072c-4e59-8c1a-4017361bf42f-kubelet-dir\") pod \"177d0275-072c-4e59-8c1a-4017361bf42f\" (UID: \"177d0275-072c-4e59-8c1a-4017361bf42f\") " Mar 17 15:41:10 crc kubenswrapper[4767]: I0317 15:41:10.232569 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/177d0275-072c-4e59-8c1a-4017361bf42f-kube-api-access\") pod \"177d0275-072c-4e59-8c1a-4017361bf42f\" (UID: \"177d0275-072c-4e59-8c1a-4017361bf42f\") " Mar 17 15:41:10 crc kubenswrapper[4767]: I0317 15:41:10.232576 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/177d0275-072c-4e59-8c1a-4017361bf42f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "177d0275-072c-4e59-8c1a-4017361bf42f" (UID: "177d0275-072c-4e59-8c1a-4017361bf42f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:41:10 crc kubenswrapper[4767]: I0317 15:41:10.232864 4767 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/177d0275-072c-4e59-8c1a-4017361bf42f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:10 crc kubenswrapper[4767]: I0317 15:41:10.240978 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/177d0275-072c-4e59-8c1a-4017361bf42f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "177d0275-072c-4e59-8c1a-4017361bf42f" (UID: "177d0275-072c-4e59-8c1a-4017361bf42f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:41:10 crc kubenswrapper[4767]: I0317 15:41:10.334394 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/177d0275-072c-4e59-8c1a-4017361bf42f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:10 crc kubenswrapper[4767]: I0317 15:41:10.945805 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"177d0275-072c-4e59-8c1a-4017361bf42f","Type":"ContainerDied","Data":"7c8f0ae6ddd5b6baca5c5c17c2c0e3118ca1dfd361b00d432293a618d6362d9a"} Mar 17 15:41:10 crc kubenswrapper[4767]: I0317 15:41:10.946095 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c8f0ae6ddd5b6baca5c5c17c2c0e3118ca1dfd361b00d432293a618d6362d9a" Mar 17 15:41:10 crc kubenswrapper[4767]: I0317 15:41:10.945828 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 15:41:13 crc kubenswrapper[4767]: E0317 15:41:12.974938 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 17 15:41:13 crc kubenswrapper[4767]: E0317 15:41:12.975437 4767 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 17 15:41:13 crc kubenswrapper[4767]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 17 15:41:13 crc kubenswrapper[4767]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wmjcq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29562700-h4fl7_openshift-infra(051f576f-ce90-4401-8944-a340dd67b274): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 17 15:41:13 crc kubenswrapper[4767]: > logger="UnhandledError" Mar 17 15:41:13 crc kubenswrapper[4767]: E0317 15:41:12.976793 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29562700-h4fl7" podUID="051f576f-ce90-4401-8944-a340dd67b274" Mar 17 15:41:13 crc kubenswrapper[4767]: E0317 15:41:13.154906 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29562700-h4fl7" podUID="051f576f-ce90-4401-8944-a340dd67b274" Mar 17 15:41:13 crc kubenswrapper[4767]: I0317 15:41:13.302031 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 15:41:18 crc kubenswrapper[4767]: I0317 15:41:18.271273 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:41:18 crc kubenswrapper[4767]: I0317 15:41:18.271792 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:41:20 crc kubenswrapper[4767]: E0317 15:41:20.683797 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 17 15:41:20 crc kubenswrapper[4767]: E0317 15:41:20.683983 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zhxhp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-79fk6_openshift-marketplace(5df92c46-85d1-4900-9916-486aa896d42a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 15:41:20 crc kubenswrapper[4767]: E0317 15:41:20.685338 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-79fk6" podUID="5df92c46-85d1-4900-9916-486aa896d42a" Mar 17 15:41:22 crc kubenswrapper[4767]: E0317 15:41:22.054891 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-79fk6" podUID="5df92c46-85d1-4900-9916-486aa896d42a" Mar 17 15:41:22 crc kubenswrapper[4767]: E0317 15:41:22.190249 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 17 15:41:22 crc kubenswrapper[4767]: E0317 15:41:22.190457 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r8lw6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-klkzj_openshift-marketplace(3dc430ad-a22a-4fe8-a4dd-5c94622da680): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 15:41:22 crc kubenswrapper[4767]: E0317 15:41:22.192396 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-klkzj" podUID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" Mar 17 15:41:23 crc kubenswrapper[4767]: E0317 15:41:23.855501 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-klkzj" podUID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" Mar 17 15:41:24 crc kubenswrapper[4767]: E0317 15:41:24.048284 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 17 15:41:24 crc kubenswrapper[4767]: E0317 15:41:24.048484 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xhfdf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zxsb9_openshift-marketplace(3f4f4f7f-0983-4551-9a1c-88c6214322f4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 15:41:24 crc kubenswrapper[4767]: E0317 15:41:24.049719 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-zxsb9" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" Mar 17 15:41:28 crc kubenswrapper[4767]: I0317 15:41:28.271544 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:41:28 crc kubenswrapper[4767]: I0317 15:41:28.274344 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:41:28 crc kubenswrapper[4767]: I0317 15:41:28.606313 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66dc66f96c-x2grn"] Mar 17 15:41:28 crc kubenswrapper[4767]: I0317 15:41:28.606520 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" podUID="f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" containerName="controller-manager" containerID="cri-o://c2ff9cf9ceadefea20ba0614ca9aa92e18891d8de41b3626a512d5f95a723dd1" gracePeriod=30 Mar 17 15:41:28 crc kubenswrapper[4767]: I0317 15:41:28.718382 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs"] Mar 17 15:41:28 crc kubenswrapper[4767]: I0317 15:41:28.718825 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" podUID="c3569d92-3c22-4a58-b535-0400b94005a2" containerName="route-controller-manager" containerID="cri-o://6e41f5befd636324890be40e1c382e04ad5a319fa271921f6130bd97e3b10b38" gracePeriod=30 Mar 17 15:41:29 crc kubenswrapper[4767]: I0317 15:41:29.505881 4767 generic.go:334] "Generic (PLEG): container finished" podID="f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" containerID="c2ff9cf9ceadefea20ba0614ca9aa92e18891d8de41b3626a512d5f95a723dd1" exitCode=0 Mar 17 15:41:29 crc kubenswrapper[4767]: I0317 15:41:29.505971 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" event={"ID":"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1","Type":"ContainerDied","Data":"c2ff9cf9ceadefea20ba0614ca9aa92e18891d8de41b3626a512d5f95a723dd1"} Mar 17 15:41:29 crc kubenswrapper[4767]: I0317 15:41:29.509736 4767 generic.go:334] "Generic (PLEG): container finished" podID="c3569d92-3c22-4a58-b535-0400b94005a2" containerID="6e41f5befd636324890be40e1c382e04ad5a319fa271921f6130bd97e3b10b38" exitCode=0 Mar 17 15:41:29 crc kubenswrapper[4767]: I0317 15:41:29.509792 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" event={"ID":"c3569d92-3c22-4a58-b535-0400b94005a2","Type":"ContainerDied","Data":"6e41f5befd636324890be40e1c382e04ad5a319fa271921f6130bd97e3b10b38"} Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.180084 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zxsb9" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.278005 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.278333 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q2dfp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-c7ktd_openshift-marketplace(f46dfeb0-7bd5-4835-a16e-8a117a3db8a3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.280198 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-c7ktd" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.280877 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.281143 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cdbm5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-vhw26_openshift-marketplace(ebb17b29-f33c-4236-bb9d-034821b17ba3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.282296 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-vhw26" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.352125 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.352913 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5lp5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-d2kvv_openshift-marketplace(7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.354278 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-d2kvv" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.389492 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.389905 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6nm8p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rq699_openshift-marketplace(7d1d71c5-161c-4ba3-96da-ed07128ff6e1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.391580 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rq699" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.451927 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.452082 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xc7fv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-q2n5r_openshift-marketplace(e76337aa-429a-49b9-99d3-e261c32ee85b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.453412 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-q2n5r" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.586535 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-59hsl" event={"ID":"471b6054-3b2a-47c9-8889-942603600da9","Type":"ContainerStarted","Data":"2b297fd0e7ea20daa56eff56749fcb564fbd15ba2e0bc336962d54eff92f4269"} Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.587680 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-59hsl" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.588532 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.588590 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.589201 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-c7ktd" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.589227 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d2kvv" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.589629 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-q2n5r" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.589686 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-vhw26" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.602163 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rq699" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.903347 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.911571 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.948220 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7"] Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.948696 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="177d0275-072c-4e59-8c1a-4017361bf42f" containerName="pruner" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.948727 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="177d0275-072c-4e59-8c1a-4017361bf42f" containerName="pruner" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.948754 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" containerName="controller-manager" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.948764 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" containerName="controller-manager" Mar 17 15:41:30 crc kubenswrapper[4767]: E0317 15:41:30.948780 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3569d92-3c22-4a58-b535-0400b94005a2" containerName="route-controller-manager" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.948792 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3569d92-3c22-4a58-b535-0400b94005a2" containerName="route-controller-manager" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.948957 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="177d0275-072c-4e59-8c1a-4017361bf42f" containerName="pruner" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.948980 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" containerName="controller-manager" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.948988 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3569d92-3c22-4a58-b535-0400b94005a2" containerName="route-controller-manager" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.949610 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.979741 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-serving-cert\") pod \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.979838 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-proxy-ca-bundles\") pod \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.979893 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-config\") pod \"c3569d92-3c22-4a58-b535-0400b94005a2\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.979933 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gb22\" (UniqueName: \"kubernetes.io/projected/c3569d92-3c22-4a58-b535-0400b94005a2-kube-api-access-4gb22\") pod \"c3569d92-3c22-4a58-b535-0400b94005a2\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.980014 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f5b7\" (UniqueName: \"kubernetes.io/projected/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-kube-api-access-8f5b7\") pod \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.980044 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-client-ca\") pod \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.980060 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-client-ca\") pod \"c3569d92-3c22-4a58-b535-0400b94005a2\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.980081 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-config\") pod \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\" (UID: \"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1\") " Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.980102 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3569d92-3c22-4a58-b535-0400b94005a2-serving-cert\") pod \"c3569d92-3c22-4a58-b535-0400b94005a2\" (UID: \"c3569d92-3c22-4a58-b535-0400b94005a2\") " Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.981110 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" (UID: "f1fc9a83-b8f5-451c-bb3a-cbb6372943c1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.982001 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-config" (OuterVolumeSpecName: "config") pod "c3569d92-3c22-4a58-b535-0400b94005a2" (UID: "c3569d92-3c22-4a58-b535-0400b94005a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.982389 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-client-ca" (OuterVolumeSpecName: "client-ca") pod "c3569d92-3c22-4a58-b535-0400b94005a2" (UID: "c3569d92-3c22-4a58-b535-0400b94005a2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.982729 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-client-ca" (OuterVolumeSpecName: "client-ca") pod "f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" (UID: "f1fc9a83-b8f5-451c-bb3a-cbb6372943c1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.982792 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-config" (OuterVolumeSpecName: "config") pod "f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" (UID: "f1fc9a83-b8f5-451c-bb3a-cbb6372943c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.989884 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-kube-api-access-8f5b7" (OuterVolumeSpecName: "kube-api-access-8f5b7") pod "f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" (UID: "f1fc9a83-b8f5-451c-bb3a-cbb6372943c1"). InnerVolumeSpecName "kube-api-access-8f5b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.990084 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3569d92-3c22-4a58-b535-0400b94005a2-kube-api-access-4gb22" (OuterVolumeSpecName: "kube-api-access-4gb22") pod "c3569d92-3c22-4a58-b535-0400b94005a2" (UID: "c3569d92-3c22-4a58-b535-0400b94005a2"). InnerVolumeSpecName "kube-api-access-4gb22". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.993781 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7"] Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.996246 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3569d92-3c22-4a58-b535-0400b94005a2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c3569d92-3c22-4a58-b535-0400b94005a2" (UID: "c3569d92-3c22-4a58-b535-0400b94005a2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:41:30 crc kubenswrapper[4767]: I0317 15:41:30.996240 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" (UID: "f1fc9a83-b8f5-451c-bb3a-cbb6372943c1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.042718 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 17 15:41:31 crc kubenswrapper[4767]: W0317 15:41:31.052451 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd49401a4_1f38_4e26_92d4_69c2085d6640.slice/crio-303663cdb320caf53b07662e4a191abf5704bfc4ceaa09559c86a55da22d9218 WatchSource:0}: Error finding container 303663cdb320caf53b07662e4a191abf5704bfc4ceaa09559c86a55da22d9218: Status 404 returned error can't find the container with id 303663cdb320caf53b07662e4a191abf5704bfc4ceaa09559c86a55da22d9218 Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081520 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-client-ca\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081660 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t4rx\" (UniqueName: \"kubernetes.io/projected/972934f3-0ca5-4093-8f03-2f6ad827685c-kube-api-access-8t4rx\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081730 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-config\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081752 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-proxy-ca-bundles\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081797 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/972934f3-0ca5-4093-8f03-2f6ad827685c-serving-cert\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081902 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081916 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081945 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gb22\" (UniqueName: \"kubernetes.io/projected/c3569d92-3c22-4a58-b535-0400b94005a2-kube-api-access-4gb22\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081955 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f5b7\" (UniqueName: \"kubernetes.io/projected/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-kube-api-access-8f5b7\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081966 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081973 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3569d92-3c22-4a58-b535-0400b94005a2-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081981 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081991 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3569d92-3c22-4a58-b535-0400b94005a2-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.081999 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.183026 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-client-ca\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.183104 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t4rx\" (UniqueName: \"kubernetes.io/projected/972934f3-0ca5-4093-8f03-2f6ad827685c-kube-api-access-8t4rx\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.183151 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-config\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.183210 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-proxy-ca-bundles\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.183243 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/972934f3-0ca5-4093-8f03-2f6ad827685c-serving-cert\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.184621 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-client-ca\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.185277 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-proxy-ca-bundles\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.185445 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-config\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.187247 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/972934f3-0ca5-4093-8f03-2f6ad827685c-serving-cert\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.199390 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t4rx\" (UniqueName: \"kubernetes.io/projected/972934f3-0ca5-4093-8f03-2f6ad827685c-kube-api-access-8t4rx\") pod \"controller-manager-7ff7b785cc-qn9k7\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.284041 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.507223 4767 csr.go:261] certificate signing request csr-sswrq is approved, waiting to be issued Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.521023 4767 csr.go:257] certificate signing request csr-sswrq is issued Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.593120 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d49401a4-1f38-4e26-92d4-69c2085d6640","Type":"ContainerStarted","Data":"ea084c55031cdaf7e3942e6d2ab50b9eeb84bdb7ec2bd283c3a8e06c0d8953f0"} Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.593210 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d49401a4-1f38-4e26-92d4-69c2085d6640","Type":"ContainerStarted","Data":"303663cdb320caf53b07662e4a191abf5704bfc4ceaa09559c86a55da22d9218"} Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.595793 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" event={"ID":"f1fc9a83-b8f5-451c-bb3a-cbb6372943c1","Type":"ContainerDied","Data":"8fe19b3bcd146dbf471e425a36f069ddb1b2b609b9cc9fbebdf009b6dee4a058"} Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.595835 4767 scope.go:117] "RemoveContainer" containerID="c2ff9cf9ceadefea20ba0614ca9aa92e18891d8de41b3626a512d5f95a723dd1" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.595850 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66dc66f96c-x2grn" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.597388 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" event={"ID":"c3569d92-3c22-4a58-b535-0400b94005a2","Type":"ContainerDied","Data":"7b41dc2703bda2c0a1f1650e31665a3976cf998ee08246863d3c310aeeadf46f"} Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.597410 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.601493 4767 generic.go:334] "Generic (PLEG): container finished" podID="051f576f-ce90-4401-8944-a340dd67b274" containerID="5b246f51d43510d489fa7052349317595bf29b0844bd0914dc6efeb056773172" exitCode=0 Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.601581 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562700-h4fl7" event={"ID":"051f576f-ce90-4401-8944-a340dd67b274","Type":"ContainerDied","Data":"5b246f51d43510d489fa7052349317595bf29b0844bd0914dc6efeb056773172"} Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.602298 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.602354 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.618485 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=44.618446085 podStartE2EDuration="44.618446085s" podCreationTimestamp="2026-03-17 15:40:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:41:31.613395407 +0000 UTC m=+283.026711484" watchObservedRunningTime="2026-03-17 15:41:31.618446085 +0000 UTC m=+283.031762142" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.620122 4767 scope.go:117] "RemoveContainer" containerID="6e41f5befd636324890be40e1c382e04ad5a319fa271921f6130bd97e3b10b38" Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.656182 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs"] Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.665603 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c545b76c8-srlrs"] Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.675186 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66dc66f96c-x2grn"] Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.687912 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-66dc66f96c-x2grn"] Mar 17 15:41:31 crc kubenswrapper[4767]: I0317 15:41:31.819421 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7"] Mar 17 15:41:31 crc kubenswrapper[4767]: W0317 15:41:31.833222 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod972934f3_0ca5_4093_8f03_2f6ad827685c.slice/crio-d71425a802db238dcb8c1e7899960b97b816c26a07474a3e4dbaf8ac298792a5 WatchSource:0}: Error finding container d71425a802db238dcb8c1e7899960b97b816c26a07474a3e4dbaf8ac298792a5: Status 404 returned error can't find the container with id d71425a802db238dcb8c1e7899960b97b816c26a07474a3e4dbaf8ac298792a5 Mar 17 15:41:32 crc kubenswrapper[4767]: I0317 15:41:32.579031 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-31 05:46:42.822432095 +0000 UTC Mar 17 15:41:32 crc kubenswrapper[4767]: I0317 15:41:32.579066 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6926h5m10.243369135s for next certificate rotation Mar 17 15:41:32 crc kubenswrapper[4767]: I0317 15:41:32.610454 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" event={"ID":"972934f3-0ca5-4093-8f03-2f6ad827685c","Type":"ContainerStarted","Data":"73d73fa69aa3ac252940026434a9f82d36c135341bb78eb1f615ffed5136be8c"} Mar 17 15:41:32 crc kubenswrapper[4767]: I0317 15:41:32.610501 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" event={"ID":"972934f3-0ca5-4093-8f03-2f6ad827685c","Type":"ContainerStarted","Data":"d71425a802db238dcb8c1e7899960b97b816c26a07474a3e4dbaf8ac298792a5"} Mar 17 15:41:32 crc kubenswrapper[4767]: I0317 15:41:32.611560 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:32 crc kubenswrapper[4767]: I0317 15:41:32.614418 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:41:32 crc kubenswrapper[4767]: I0317 15:41:32.614452 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:41:32 crc kubenswrapper[4767]: I0317 15:41:32.620944 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:32 crc kubenswrapper[4767]: I0317 15:41:32.629898 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" podStartSLOduration=4.62987428 podStartE2EDuration="4.62987428s" podCreationTimestamp="2026-03-17 15:41:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:41:32.626067398 +0000 UTC m=+284.039383455" watchObservedRunningTime="2026-03-17 15:41:32.62987428 +0000 UTC m=+284.043190337" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.214293 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562700-h4fl7" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.274466 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2"] Mar 17 15:41:33 crc kubenswrapper[4767]: E0317 15:41:33.274836 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="051f576f-ce90-4401-8944-a340dd67b274" containerName="oc" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.274857 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="051f576f-ce90-4401-8944-a340dd67b274" containerName="oc" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.275004 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="051f576f-ce90-4401-8944-a340dd67b274" containerName="oc" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.275639 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.285035 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.285134 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.285503 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.285689 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.285859 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.286044 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.301890 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2"] Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.367471 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3569d92-3c22-4a58-b535-0400b94005a2" path="/var/lib/kubelet/pods/c3569d92-3c22-4a58-b535-0400b94005a2/volumes" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.368078 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1fc9a83-b8f5-451c-bb3a-cbb6372943c1" path="/var/lib/kubelet/pods/f1fc9a83-b8f5-451c-bb3a-cbb6372943c1/volumes" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.390624 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmjcq\" (UniqueName: \"kubernetes.io/projected/051f576f-ce90-4401-8944-a340dd67b274-kube-api-access-wmjcq\") pod \"051f576f-ce90-4401-8944-a340dd67b274\" (UID: \"051f576f-ce90-4401-8944-a340dd67b274\") " Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.391272 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67b25981-18f2-4dd7-9f1f-9d67c86ca634-serving-cert\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.391668 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-config\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.392351 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-client-ca\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.392583 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvnvk\" (UniqueName: \"kubernetes.io/projected/67b25981-18f2-4dd7-9f1f-9d67c86ca634-kube-api-access-hvnvk\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.397203 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/051f576f-ce90-4401-8944-a340dd67b274-kube-api-access-wmjcq" (OuterVolumeSpecName: "kube-api-access-wmjcq") pod "051f576f-ce90-4401-8944-a340dd67b274" (UID: "051f576f-ce90-4401-8944-a340dd67b274"). InnerVolumeSpecName "kube-api-access-wmjcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.494500 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67b25981-18f2-4dd7-9f1f-9d67c86ca634-serving-cert\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.495529 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-config\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.495567 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-client-ca\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.495614 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvnvk\" (UniqueName: \"kubernetes.io/projected/67b25981-18f2-4dd7-9f1f-9d67c86ca634-kube-api-access-hvnvk\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.495680 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmjcq\" (UniqueName: \"kubernetes.io/projected/051f576f-ce90-4401-8944-a340dd67b274-kube-api-access-wmjcq\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.497327 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-client-ca\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.498194 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-config\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.499591 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67b25981-18f2-4dd7-9f1f-9d67c86ca634-serving-cert\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.516157 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvnvk\" (UniqueName: \"kubernetes.io/projected/67b25981-18f2-4dd7-9f1f-9d67c86ca634-kube-api-access-hvnvk\") pod \"route-controller-manager-74cbc5ff48-wdjj2\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.579932 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-26 21:39:29.737547685 +0000 UTC Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.579994 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6821h57m56.157557537s for next certificate rotation Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.608778 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.623475 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562700-h4fl7" event={"ID":"051f576f-ce90-4401-8944-a340dd67b274","Type":"ContainerDied","Data":"d57d73aa046e2e0fbbb9d78f1e72bcf4005f9031f7daf199e4c4411a2828f5fa"} Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.623937 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d57d73aa046e2e0fbbb9d78f1e72bcf4005f9031f7daf199e4c4411a2828f5fa" Mar 17 15:41:33 crc kubenswrapper[4767]: I0317 15:41:33.624103 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562700-h4fl7" Mar 17 15:41:34 crc kubenswrapper[4767]: I0317 15:41:34.167252 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:41:34 crc kubenswrapper[4767]: I0317 15:41:34.167924 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:41:34 crc kubenswrapper[4767]: I0317 15:41:34.168021 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:41:34 crc kubenswrapper[4767]: I0317 15:41:34.169283 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 15:41:34 crc kubenswrapper[4767]: I0317 15:41:34.169391 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b" gracePeriod=600 Mar 17 15:41:34 crc kubenswrapper[4767]: I0317 15:41:34.256828 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2"] Mar 17 15:41:34 crc kubenswrapper[4767]: W0317 15:41:34.259354 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67b25981_18f2_4dd7_9f1f_9d67c86ca634.slice/crio-cedf23a1cf3846fff101bf18659699f7a508e9685126decf5f004732e16bae9c WatchSource:0}: Error finding container cedf23a1cf3846fff101bf18659699f7a508e9685126decf5f004732e16bae9c: Status 404 returned error can't find the container with id cedf23a1cf3846fff101bf18659699f7a508e9685126decf5f004732e16bae9c Mar 17 15:41:34 crc kubenswrapper[4767]: I0317 15:41:34.629590 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" event={"ID":"67b25981-18f2-4dd7-9f1f-9d67c86ca634","Type":"ContainerStarted","Data":"cedf23a1cf3846fff101bf18659699f7a508e9685126decf5f004732e16bae9c"} Mar 17 15:41:35 crc kubenswrapper[4767]: I0317 15:41:35.637974 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" event={"ID":"67b25981-18f2-4dd7-9f1f-9d67c86ca634","Type":"ContainerStarted","Data":"ec69d7ba77cf396fe7787d0cc70f8b96ec3d2132591c80d17275bee05b24eb0c"} Mar 17 15:41:35 crc kubenswrapper[4767]: I0317 15:41:35.639170 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:35 crc kubenswrapper[4767]: I0317 15:41:35.648792 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:35 crc kubenswrapper[4767]: I0317 15:41:35.649225 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b" exitCode=0 Mar 17 15:41:35 crc kubenswrapper[4767]: I0317 15:41:35.649276 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b"} Mar 17 15:41:35 crc kubenswrapper[4767]: I0317 15:41:35.649309 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"3c904b546ac42362d6e52f6a82d664f6ec63ec5a64d0aa1a9ae36b3788164828"} Mar 17 15:41:35 crc kubenswrapper[4767]: I0317 15:41:35.668798 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" podStartSLOduration=7.668769162 podStartE2EDuration="7.668769162s" podCreationTimestamp="2026-03-17 15:41:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:41:35.661387176 +0000 UTC m=+287.074703263" watchObservedRunningTime="2026-03-17 15:41:35.668769162 +0000 UTC m=+287.082085209" Mar 17 15:41:36 crc kubenswrapper[4767]: E0317 15:41:36.138327 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5df92c46_85d1_4900_9916_486aa896d42a.slice/crio-d1fd9a1e142d352f554f801a97facb4f707993dab7058ae4fb055e177e36e3eb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5df92c46_85d1_4900_9916_486aa896d42a.slice/crio-conmon-d1fd9a1e142d352f554f801a97facb4f707993dab7058ae4fb055e177e36e3eb.scope\": RecentStats: unable to find data in memory cache]" Mar 17 15:41:36 crc kubenswrapper[4767]: I0317 15:41:36.948476 4767 generic.go:334] "Generic (PLEG): container finished" podID="5df92c46-85d1-4900-9916-486aa896d42a" containerID="d1fd9a1e142d352f554f801a97facb4f707993dab7058ae4fb055e177e36e3eb" exitCode=0 Mar 17 15:41:36 crc kubenswrapper[4767]: I0317 15:41:36.948588 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79fk6" event={"ID":"5df92c46-85d1-4900-9916-486aa896d42a","Type":"ContainerDied","Data":"d1fd9a1e142d352f554f801a97facb4f707993dab7058ae4fb055e177e36e3eb"} Mar 17 15:41:38 crc kubenswrapper[4767]: I0317 15:41:38.078152 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79fk6" event={"ID":"5df92c46-85d1-4900-9916-486aa896d42a","Type":"ContainerStarted","Data":"901c54ac466d898a9166d8e9a04549bfe43ab282abc9ab76a629b42e6ad55297"} Mar 17 15:41:38 crc kubenswrapper[4767]: I0317 15:41:38.105226 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-79fk6" podStartSLOduration=9.521241533 podStartE2EDuration="1m18.105209451s" podCreationTimestamp="2026-03-17 15:40:20 +0000 UTC" firstStartedPulling="2026-03-17 15:40:28.834987731 +0000 UTC m=+220.248303778" lastFinishedPulling="2026-03-17 15:41:37.418955659 +0000 UTC m=+288.832271696" observedRunningTime="2026-03-17 15:41:38.099761851 +0000 UTC m=+289.513077918" watchObservedRunningTime="2026-03-17 15:41:38.105209451 +0000 UTC m=+289.518525498" Mar 17 15:41:38 crc kubenswrapper[4767]: I0317 15:41:38.270790 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:41:38 crc kubenswrapper[4767]: I0317 15:41:38.271423 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:41:38 crc kubenswrapper[4767]: I0317 15:41:38.271522 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:41:38 crc kubenswrapper[4767]: I0317 15:41:38.272040 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:41:42 crc kubenswrapper[4767]: I0317 15:41:42.054291 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:41:42 crc kubenswrapper[4767]: I0317 15:41:42.057040 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:41:43 crc kubenswrapper[4767]: I0317 15:41:43.461237 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:41:43 crc kubenswrapper[4767]: I0317 15:41:43.519147 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:41:44 crc kubenswrapper[4767]: I0317 15:41:44.156287 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkzj" event={"ID":"3dc430ad-a22a-4fe8-a4dd-5c94622da680","Type":"ContainerStarted","Data":"42f59dbaacd0e780dee973ffc979cd875439c6bb96338a5ce64ed135ea0ee59d"} Mar 17 15:41:45 crc kubenswrapper[4767]: I0317 15:41:45.050648 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-79fk6"] Mar 17 15:41:45 crc kubenswrapper[4767]: I0317 15:41:45.304205 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-79fk6" podUID="5df92c46-85d1-4900-9916-486aa896d42a" containerName="registry-server" containerID="cri-o://901c54ac466d898a9166d8e9a04549bfe43ab282abc9ab76a629b42e6ad55297" gracePeriod=2 Mar 17 15:41:46 crc kubenswrapper[4767]: I0317 15:41:46.314512 4767 generic.go:334] "Generic (PLEG): container finished" podID="5df92c46-85d1-4900-9916-486aa896d42a" containerID="901c54ac466d898a9166d8e9a04549bfe43ab282abc9ab76a629b42e6ad55297" exitCode=0 Mar 17 15:41:46 crc kubenswrapper[4767]: I0317 15:41:46.314592 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79fk6" event={"ID":"5df92c46-85d1-4900-9916-486aa896d42a","Type":"ContainerDied","Data":"901c54ac466d898a9166d8e9a04549bfe43ab282abc9ab76a629b42e6ad55297"} Mar 17 15:41:46 crc kubenswrapper[4767]: I0317 15:41:46.317499 4767 generic.go:334] "Generic (PLEG): container finished" podID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" containerID="42f59dbaacd0e780dee973ffc979cd875439c6bb96338a5ce64ed135ea0ee59d" exitCode=0 Mar 17 15:41:46 crc kubenswrapper[4767]: I0317 15:41:46.317562 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkzj" event={"ID":"3dc430ad-a22a-4fe8-a4dd-5c94622da680","Type":"ContainerDied","Data":"42f59dbaacd0e780dee973ffc979cd875439c6bb96338a5ce64ed135ea0ee59d"} Mar 17 15:41:46 crc kubenswrapper[4767]: I0317 15:41:46.319337 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2n5r" event={"ID":"e76337aa-429a-49b9-99d3-e261c32ee85b","Type":"ContainerStarted","Data":"8c0e4705b915b89f591cb5140d524eda895c66c18c99ed647317cb1fc9d4d3b3"} Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.271212 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.271626 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.272319 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.272504 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.506560 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.652526 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7"] Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.652806 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" podUID="972934f3-0ca5-4093-8f03-2f6ad827685c" containerName="controller-manager" containerID="cri-o://73d73fa69aa3ac252940026434a9f82d36c135341bb78eb1f615ffed5136be8c" gracePeriod=30 Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.665034 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhxhp\" (UniqueName: \"kubernetes.io/projected/5df92c46-85d1-4900-9916-486aa896d42a-kube-api-access-zhxhp\") pod \"5df92c46-85d1-4900-9916-486aa896d42a\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.665130 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-catalog-content\") pod \"5df92c46-85d1-4900-9916-486aa896d42a\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.665209 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-utilities\") pod \"5df92c46-85d1-4900-9916-486aa896d42a\" (UID: \"5df92c46-85d1-4900-9916-486aa896d42a\") " Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.666632 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-utilities" (OuterVolumeSpecName: "utilities") pod "5df92c46-85d1-4900-9916-486aa896d42a" (UID: "5df92c46-85d1-4900-9916-486aa896d42a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.680279 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5df92c46-85d1-4900-9916-486aa896d42a-kube-api-access-zhxhp" (OuterVolumeSpecName: "kube-api-access-zhxhp") pod "5df92c46-85d1-4900-9916-486aa896d42a" (UID: "5df92c46-85d1-4900-9916-486aa896d42a"). InnerVolumeSpecName "kube-api-access-zhxhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.689530 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2"] Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.689790 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" podUID="67b25981-18f2-4dd7-9f1f-9d67c86ca634" containerName="route-controller-manager" containerID="cri-o://ec69d7ba77cf396fe7787d0cc70f8b96ec3d2132591c80d17275bee05b24eb0c" gracePeriod=30 Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.766746 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhxhp\" (UniqueName: \"kubernetes.io/projected/5df92c46-85d1-4900-9916-486aa896d42a-kube-api-access-zhxhp\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.766803 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.896085 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5df92c46-85d1-4900-9916-486aa896d42a" (UID: "5df92c46-85d1-4900-9916-486aa896d42a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:41:48 crc kubenswrapper[4767]: I0317 15:41:48.969611 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df92c46-85d1-4900-9916-486aa896d42a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:49 crc kubenswrapper[4767]: I0317 15:41:49.356839 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-79fk6" Mar 17 15:41:49 crc kubenswrapper[4767]: I0317 15:41:49.366288 4767 generic.go:334] "Generic (PLEG): container finished" podID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerID="8c0e4705b915b89f591cb5140d524eda895c66c18c99ed647317cb1fc9d4d3b3" exitCode=0 Mar 17 15:41:49 crc kubenswrapper[4767]: I0317 15:41:49.373808 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79fk6" event={"ID":"5df92c46-85d1-4900-9916-486aa896d42a","Type":"ContainerDied","Data":"fb70c784f417fb6c4726bfcd62f807ec81c6f6f9300e3ea14facd1c59d2bdb62"} Mar 17 15:41:49 crc kubenswrapper[4767]: I0317 15:41:49.373845 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2n5r" event={"ID":"e76337aa-429a-49b9-99d3-e261c32ee85b","Type":"ContainerDied","Data":"8c0e4705b915b89f591cb5140d524eda895c66c18c99ed647317cb1fc9d4d3b3"} Mar 17 15:41:49 crc kubenswrapper[4767]: I0317 15:41:49.373921 4767 scope.go:117] "RemoveContainer" containerID="901c54ac466d898a9166d8e9a04549bfe43ab282abc9ab76a629b42e6ad55297" Mar 17 15:41:49 crc kubenswrapper[4767]: I0317 15:41:49.482723 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-79fk6"] Mar 17 15:41:49 crc kubenswrapper[4767]: I0317 15:41:49.488163 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-79fk6"] Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.375253 4767 generic.go:334] "Generic (PLEG): container finished" podID="972934f3-0ca5-4093-8f03-2f6ad827685c" containerID="73d73fa69aa3ac252940026434a9f82d36c135341bb78eb1f615ffed5136be8c" exitCode=0 Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.375351 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" event={"ID":"972934f3-0ca5-4093-8f03-2f6ad827685c","Type":"ContainerDied","Data":"73d73fa69aa3ac252940026434a9f82d36c135341bb78eb1f615ffed5136be8c"} Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.376888 4767 generic.go:334] "Generic (PLEG): container finished" podID="67b25981-18f2-4dd7-9f1f-9d67c86ca634" containerID="ec69d7ba77cf396fe7787d0cc70f8b96ec3d2132591c80d17275bee05b24eb0c" exitCode=0 Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.376934 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" event={"ID":"67b25981-18f2-4dd7-9f1f-9d67c86ca634","Type":"ContainerDied","Data":"ec69d7ba77cf396fe7787d0cc70f8b96ec3d2132591c80d17275bee05b24eb0c"} Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.874548 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.879509 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.907580 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw"] Mar 17 15:41:50 crc kubenswrapper[4767]: E0317 15:41:50.907949 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67b25981-18f2-4dd7-9f1f-9d67c86ca634" containerName="route-controller-manager" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.907979 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="67b25981-18f2-4dd7-9f1f-9d67c86ca634" containerName="route-controller-manager" Mar 17 15:41:50 crc kubenswrapper[4767]: E0317 15:41:50.907997 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="972934f3-0ca5-4093-8f03-2f6ad827685c" containerName="controller-manager" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.908007 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="972934f3-0ca5-4093-8f03-2f6ad827685c" containerName="controller-manager" Mar 17 15:41:50 crc kubenswrapper[4767]: E0317 15:41:50.908021 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5df92c46-85d1-4900-9916-486aa896d42a" containerName="extract-content" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.908029 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5df92c46-85d1-4900-9916-486aa896d42a" containerName="extract-content" Mar 17 15:41:50 crc kubenswrapper[4767]: E0317 15:41:50.908042 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5df92c46-85d1-4900-9916-486aa896d42a" containerName="registry-server" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.908050 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5df92c46-85d1-4900-9916-486aa896d42a" containerName="registry-server" Mar 17 15:41:50 crc kubenswrapper[4767]: E0317 15:41:50.908065 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5df92c46-85d1-4900-9916-486aa896d42a" containerName="extract-utilities" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.908073 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5df92c46-85d1-4900-9916-486aa896d42a" containerName="extract-utilities" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.908240 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5df92c46-85d1-4900-9916-486aa896d42a" containerName="registry-server" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.908261 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="972934f3-0ca5-4093-8f03-2f6ad827685c" containerName="controller-manager" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.908278 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="67b25981-18f2-4dd7-9f1f-9d67c86ca634" containerName="route-controller-manager" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.908803 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:50 crc kubenswrapper[4767]: I0317 15:41:50.921657 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw"] Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.004236 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-config\") pod \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.004398 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-proxy-ca-bundles\") pod \"972934f3-0ca5-4093-8f03-2f6ad827685c\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.004483 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/972934f3-0ca5-4093-8f03-2f6ad827685c-serving-cert\") pod \"972934f3-0ca5-4093-8f03-2f6ad827685c\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.004518 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-client-ca\") pod \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.004617 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-config\") pod \"972934f3-0ca5-4093-8f03-2f6ad827685c\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.004667 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67b25981-18f2-4dd7-9f1f-9d67c86ca634-serving-cert\") pod \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.004783 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t4rx\" (UniqueName: \"kubernetes.io/projected/972934f3-0ca5-4093-8f03-2f6ad827685c-kube-api-access-8t4rx\") pod \"972934f3-0ca5-4093-8f03-2f6ad827685c\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.004820 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvnvk\" (UniqueName: \"kubernetes.io/projected/67b25981-18f2-4dd7-9f1f-9d67c86ca634-kube-api-access-hvnvk\") pod \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\" (UID: \"67b25981-18f2-4dd7-9f1f-9d67c86ca634\") " Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.004858 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-client-ca\") pod \"972934f3-0ca5-4093-8f03-2f6ad827685c\" (UID: \"972934f3-0ca5-4093-8f03-2f6ad827685c\") " Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.005125 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e548a4-cb50-4864-ba89-e33e8a3dda4e-serving-cert\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.005140 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-config" (OuterVolumeSpecName: "config") pod "67b25981-18f2-4dd7-9f1f-9d67c86ca634" (UID: "67b25981-18f2-4dd7-9f1f-9d67c86ca634"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.005269 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-config\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.005301 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-client-ca\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.005350 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jfsc\" (UniqueName: \"kubernetes.io/projected/06e548a4-cb50-4864-ba89-e33e8a3dda4e-kube-api-access-5jfsc\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.005392 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.005884 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "972934f3-0ca5-4093-8f03-2f6ad827685c" (UID: "972934f3-0ca5-4093-8f03-2f6ad827685c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.006116 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-client-ca" (OuterVolumeSpecName: "client-ca") pod "972934f3-0ca5-4093-8f03-2f6ad827685c" (UID: "972934f3-0ca5-4093-8f03-2f6ad827685c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.006418 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-client-ca" (OuterVolumeSpecName: "client-ca") pod "67b25981-18f2-4dd7-9f1f-9d67c86ca634" (UID: "67b25981-18f2-4dd7-9f1f-9d67c86ca634"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.006719 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-config" (OuterVolumeSpecName: "config") pod "972934f3-0ca5-4093-8f03-2f6ad827685c" (UID: "972934f3-0ca5-4093-8f03-2f6ad827685c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.009772 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/972934f3-0ca5-4093-8f03-2f6ad827685c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "972934f3-0ca5-4093-8f03-2f6ad827685c" (UID: "972934f3-0ca5-4093-8f03-2f6ad827685c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.010112 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b25981-18f2-4dd7-9f1f-9d67c86ca634-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "67b25981-18f2-4dd7-9f1f-9d67c86ca634" (UID: "67b25981-18f2-4dd7-9f1f-9d67c86ca634"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.010869 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67b25981-18f2-4dd7-9f1f-9d67c86ca634-kube-api-access-hvnvk" (OuterVolumeSpecName: "kube-api-access-hvnvk") pod "67b25981-18f2-4dd7-9f1f-9d67c86ca634" (UID: "67b25981-18f2-4dd7-9f1f-9d67c86ca634"). InnerVolumeSpecName "kube-api-access-hvnvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.011727 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/972934f3-0ca5-4093-8f03-2f6ad827685c-kube-api-access-8t4rx" (OuterVolumeSpecName: "kube-api-access-8t4rx") pod "972934f3-0ca5-4093-8f03-2f6ad827685c" (UID: "972934f3-0ca5-4093-8f03-2f6ad827685c"). InnerVolumeSpecName "kube-api-access-8t4rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106516 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-config\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106608 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-client-ca\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106653 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jfsc\" (UniqueName: \"kubernetes.io/projected/06e548a4-cb50-4864-ba89-e33e8a3dda4e-kube-api-access-5jfsc\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106723 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e548a4-cb50-4864-ba89-e33e8a3dda4e-serving-cert\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106810 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t4rx\" (UniqueName: \"kubernetes.io/projected/972934f3-0ca5-4093-8f03-2f6ad827685c-kube-api-access-8t4rx\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106827 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvnvk\" (UniqueName: \"kubernetes.io/projected/67b25981-18f2-4dd7-9f1f-9d67c86ca634-kube-api-access-hvnvk\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106845 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106857 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106868 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/972934f3-0ca5-4093-8f03-2f6ad827685c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106883 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67b25981-18f2-4dd7-9f1f-9d67c86ca634-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106896 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/972934f3-0ca5-4093-8f03-2f6ad827685c-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.106909 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67b25981-18f2-4dd7-9f1f-9d67c86ca634-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.107931 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-client-ca\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.107985 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-config\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.111990 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e548a4-cb50-4864-ba89-e33e8a3dda4e-serving-cert\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.123232 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jfsc\" (UniqueName: \"kubernetes.io/projected/06e548a4-cb50-4864-ba89-e33e8a3dda4e-kube-api-access-5jfsc\") pod \"route-controller-manager-7697ffcff9-hbkkw\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.226523 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.233632 4767 scope.go:117] "RemoveContainer" containerID="d1fd9a1e142d352f554f801a97facb4f707993dab7058ae4fb055e177e36e3eb" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.362261 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5df92c46-85d1-4900-9916-486aa896d42a" path="/var/lib/kubelet/pods/5df92c46-85d1-4900-9916-486aa896d42a/volumes" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.401623 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.402329 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7" event={"ID":"972934f3-0ca5-4093-8f03-2f6ad827685c","Type":"ContainerDied","Data":"d71425a802db238dcb8c1e7899960b97b816c26a07474a3e4dbaf8ac298792a5"} Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.407015 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" event={"ID":"67b25981-18f2-4dd7-9f1f-9d67c86ca634","Type":"ContainerDied","Data":"cedf23a1cf3846fff101bf18659699f7a508e9685126decf5f004732e16bae9c"} Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.407108 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2" Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.420906 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7"] Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.424735 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7ff7b785cc-qn9k7"] Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.437993 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2"] Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.441721 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74cbc5ff48-wdjj2"] Mar 17 15:41:51 crc kubenswrapper[4767]: I0317 15:41:51.825503 4767 scope.go:117] "RemoveContainer" containerID="ff528ef531b732270c2a7159eabb44c8387e17ced0fcef2301dd1798a7cac3d3" Mar 17 15:41:52 crc kubenswrapper[4767]: I0317 15:41:52.062707 4767 scope.go:117] "RemoveContainer" containerID="73d73fa69aa3ac252940026434a9f82d36c135341bb78eb1f615ffed5136be8c" Mar 17 15:41:52 crc kubenswrapper[4767]: I0317 15:41:52.106025 4767 scope.go:117] "RemoveContainer" containerID="ec69d7ba77cf396fe7787d0cc70f8b96ec3d2132591c80d17275bee05b24eb0c" Mar 17 15:41:52 crc kubenswrapper[4767]: I0317 15:41:52.803385 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw"] Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.587032 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67b25981-18f2-4dd7-9f1f-9d67c86ca634" path="/var/lib/kubelet/pods/67b25981-18f2-4dd7-9f1f-9d67c86ca634/volumes" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.589543 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="972934f3-0ca5-4093-8f03-2f6ad827685c" path="/var/lib/kubelet/pods/972934f3-0ca5-4093-8f03-2f6ad827685c/volumes" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.590243 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhw26" event={"ID":"ebb17b29-f33c-4236-bb9d-034821b17ba3","Type":"ContainerStarted","Data":"e76f0e7f83dec65bda69231c082bc4ce8cccc7fa902c2933d856c707983c139f"} Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.590291 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6b96464484-gwtw5"] Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.591083 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b96464484-gwtw5"] Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.591613 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.594836 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.595265 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7ktd" event={"ID":"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3","Type":"ContainerStarted","Data":"74224c64630e0da55b91e6ac81b9b9b624e553c10f72323f0c1de1748eab9c0a"} Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.596064 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq699" event={"ID":"7d1d71c5-161c-4ba3-96da-ed07128ff6e1","Type":"ContainerStarted","Data":"17c61359073c5df3cd11758b35ca04cc055d0061fcfd6b7c625bbe56ddc4736a"} Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.597158 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.598327 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.598722 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkzj" event={"ID":"3dc430ad-a22a-4fe8-a4dd-5c94622da680","Type":"ContainerStarted","Data":"b2b986e0261578449e1f42241f3d405db374fb6e7e52966e8a4003ce254adbf3"} Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.599556 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.599643 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.599787 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.600945 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxsb9" event={"ID":"3f4f4f7f-0983-4551-9a1c-88c6214322f4","Type":"ContainerStarted","Data":"83cf83ceebcb7c8c1ed7e69e335c1a5d086be3dc6d45ff7f0b4889b99c79e99c"} Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.606325 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.616593 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2n5r" event={"ID":"e76337aa-429a-49b9-99d3-e261c32ee85b","Type":"ContainerStarted","Data":"e90359a0dfe7155779e0ad33fa338d37c63ebe2059107a4279d83d66ee56158b"} Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.636642 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" event={"ID":"06e548a4-cb50-4864-ba89-e33e8a3dda4e","Type":"ContainerStarted","Data":"640bd0bd4c9604176ebf62f8e39d92793b2859c6056c7bcc42388a6e73e7e6bb"} Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.640761 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2kvv" event={"ID":"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a","Type":"ContainerStarted","Data":"848997677f41edf0cfe605cd77c42edab072bab43d1fd9876ee6557f9f57c328"} Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.755214 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-klkzj" podStartSLOduration=9.290333945 podStartE2EDuration="1m33.755194009s" podCreationTimestamp="2026-03-17 15:40:20 +0000 UTC" firstStartedPulling="2026-03-17 15:40:27.36147291 +0000 UTC m=+218.774788957" lastFinishedPulling="2026-03-17 15:41:51.826332974 +0000 UTC m=+303.239649021" observedRunningTime="2026-03-17 15:41:53.754158018 +0000 UTC m=+305.167474065" watchObservedRunningTime="2026-03-17 15:41:53.755194009 +0000 UTC m=+305.168510086" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.991521 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g9tx\" (UniqueName: \"kubernetes.io/projected/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-kube-api-access-2g9tx\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.991781 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-proxy-ca-bundles\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.991896 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-config\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.992089 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-serving-cert\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:53 crc kubenswrapper[4767]: I0317 15:41:53.992286 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-client-ca\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.055856 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q2n5r" podStartSLOduration=7.343881507 podStartE2EDuration="1m31.055836584s" podCreationTimestamp="2026-03-17 15:40:23 +0000 UTC" firstStartedPulling="2026-03-17 15:40:28.772979967 +0000 UTC m=+220.186296004" lastFinishedPulling="2026-03-17 15:41:52.484935024 +0000 UTC m=+303.898251081" observedRunningTime="2026-03-17 15:41:54.05058045 +0000 UTC m=+305.463896517" watchObservedRunningTime="2026-03-17 15:41:54.055836584 +0000 UTC m=+305.469152631" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.093465 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-config\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.093764 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-serving-cert\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.093866 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-client-ca\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.094001 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g9tx\" (UniqueName: \"kubernetes.io/projected/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-kube-api-access-2g9tx\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.094117 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-proxy-ca-bundles\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.099048 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-client-ca\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.100235 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-config\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.108008 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-serving-cert\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.118812 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-proxy-ca-bundles\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.131172 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g9tx\" (UniqueName: \"kubernetes.io/projected/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-kube-api-access-2g9tx\") pod \"controller-manager-6b96464484-gwtw5\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.213562 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.740449 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" event={"ID":"06e548a4-cb50-4864-ba89-e33e8a3dda4e","Type":"ContainerStarted","Data":"aea7085a2c9a5f6e02a18c90d1bc6249ac3cc93e159b69e1c4ef5bbab41925e6"} Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.744863 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.832051 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:41:54 crc kubenswrapper[4767]: I0317 15:41:54.891072 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" podStartSLOduration=6.891049823 podStartE2EDuration="6.891049823s" podCreationTimestamp="2026-03-17 15:41:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:41:54.781374777 +0000 UTC m=+306.194690844" watchObservedRunningTime="2026-03-17 15:41:54.891049823 +0000 UTC m=+306.304365880" Mar 17 15:41:55 crc kubenswrapper[4767]: I0317 15:41:55.192558 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:41:55 crc kubenswrapper[4767]: I0317 15:41:55.192599 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:41:56 crc kubenswrapper[4767]: I0317 15:41:55.818439 4767 generic.go:334] "Generic (PLEG): container finished" podID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" containerID="83cf83ceebcb7c8c1ed7e69e335c1a5d086be3dc6d45ff7f0b4889b99c79e99c" exitCode=0 Mar 17 15:41:56 crc kubenswrapper[4767]: I0317 15:41:55.820113 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxsb9" event={"ID":"3f4f4f7f-0983-4551-9a1c-88c6214322f4","Type":"ContainerDied","Data":"83cf83ceebcb7c8c1ed7e69e335c1a5d086be3dc6d45ff7f0b4889b99c79e99c"} Mar 17 15:41:56 crc kubenswrapper[4767]: I0317 15:41:56.049725 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b96464484-gwtw5"] Mar 17 15:41:56 crc kubenswrapper[4767]: W0317 15:41:56.134349 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9bbbb6f_72aa_4d0d_809b_566c9f176d6b.slice/crio-1ac61668b742c00028227095883917a88f4e8ad912716050d8e27710ab880554 WatchSource:0}: Error finding container 1ac61668b742c00028227095883917a88f4e8ad912716050d8e27710ab880554: Status 404 returned error can't find the container with id 1ac61668b742c00028227095883917a88f4e8ad912716050d8e27710ab880554 Mar 17 15:41:56 crc kubenswrapper[4767]: I0317 15:41:56.418110 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q2n5r" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerName="registry-server" probeResult="failure" output=< Mar 17 15:41:56 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 15:41:56 crc kubenswrapper[4767]: > Mar 17 15:41:56 crc kubenswrapper[4767]: I0317 15:41:56.824439 4767 generic.go:334] "Generic (PLEG): container finished" podID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" containerID="848997677f41edf0cfe605cd77c42edab072bab43d1fd9876ee6557f9f57c328" exitCode=0 Mar 17 15:41:56 crc kubenswrapper[4767]: I0317 15:41:56.824721 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2kvv" event={"ID":"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a","Type":"ContainerDied","Data":"848997677f41edf0cfe605cd77c42edab072bab43d1fd9876ee6557f9f57c328"} Mar 17 15:41:56 crc kubenswrapper[4767]: I0317 15:41:56.828333 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" event={"ID":"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b","Type":"ContainerStarted","Data":"1ac61668b742c00028227095883917a88f4e8ad912716050d8e27710ab880554"} Mar 17 15:41:56 crc kubenswrapper[4767]: I0317 15:41:56.832112 4767 generic.go:334] "Generic (PLEG): container finished" podID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerID="74224c64630e0da55b91e6ac81b9b9b624e553c10f72323f0c1de1748eab9c0a" exitCode=0 Mar 17 15:41:56 crc kubenswrapper[4767]: I0317 15:41:56.832200 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7ktd" event={"ID":"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3","Type":"ContainerDied","Data":"74224c64630e0da55b91e6ac81b9b9b624e553c10f72323f0c1de1748eab9c0a"} Mar 17 15:41:56 crc kubenswrapper[4767]: I0317 15:41:56.839418 4767 generic.go:334] "Generic (PLEG): container finished" podID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerID="17c61359073c5df3cd11758b35ca04cc055d0061fcfd6b7c625bbe56ddc4736a" exitCode=0 Mar 17 15:41:56 crc kubenswrapper[4767]: I0317 15:41:56.840626 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq699" event={"ID":"7d1d71c5-161c-4ba3-96da-ed07128ff6e1","Type":"ContainerDied","Data":"17c61359073c5df3cd11758b35ca04cc055d0061fcfd6b7c625bbe56ddc4736a"} Mar 17 15:41:57 crc kubenswrapper[4767]: I0317 15:41:57.897029 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxsb9" event={"ID":"3f4f4f7f-0983-4551-9a1c-88c6214322f4","Type":"ContainerStarted","Data":"0a1bddc789e7b1c9b052b88d48562c8043c6d450b5ff48b838b7c57ed7d25024"} Mar 17 15:41:57 crc kubenswrapper[4767]: I0317 15:41:57.899633 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" event={"ID":"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b","Type":"ContainerStarted","Data":"f5bc9e45c59aa87b62bc954daf73e543ba9fb8e158d07ceffe6d501b83a0c856"} Mar 17 15:41:57 crc kubenswrapper[4767]: I0317 15:41:57.899986 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:57 crc kubenswrapper[4767]: I0317 15:41:57.912324 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:41:57 crc kubenswrapper[4767]: I0317 15:41:57.943931 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zxsb9" podStartSLOduration=7.325614715 podStartE2EDuration="1m34.943906824s" podCreationTimestamp="2026-03-17 15:40:23 +0000 UTC" firstStartedPulling="2026-03-17 15:40:28.734019599 +0000 UTC m=+220.147335646" lastFinishedPulling="2026-03-17 15:41:56.352311688 +0000 UTC m=+307.765627755" observedRunningTime="2026-03-17 15:41:57.923735553 +0000 UTC m=+309.337051620" watchObservedRunningTime="2026-03-17 15:41:57.943906824 +0000 UTC m=+309.357222881" Mar 17 15:41:58 crc kubenswrapper[4767]: I0317 15:41:58.307453 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-59hsl" Mar 17 15:41:58 crc kubenswrapper[4767]: I0317 15:41:58.331393 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" podStartSLOduration=10.331340394 podStartE2EDuration="10.331340394s" podCreationTimestamp="2026-03-17 15:41:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:41:57.94544953 +0000 UTC m=+309.358765597" watchObservedRunningTime="2026-03-17 15:41:58.331340394 +0000 UTC m=+309.744656441" Mar 17 15:41:59 crc kubenswrapper[4767]: I0317 15:41:58.957774 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7ktd" event={"ID":"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3","Type":"ContainerStarted","Data":"81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208"} Mar 17 15:41:59 crc kubenswrapper[4767]: I0317 15:41:58.995848 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c7ktd" podStartSLOduration=7.312109547 podStartE2EDuration="1m35.995818118s" podCreationTimestamp="2026-03-17 15:40:23 +0000 UTC" firstStartedPulling="2026-03-17 15:40:28.72668891 +0000 UTC m=+220.140004957" lastFinishedPulling="2026-03-17 15:41:57.410397471 +0000 UTC m=+308.823713528" observedRunningTime="2026-03-17 15:41:58.983905168 +0000 UTC m=+310.397221245" watchObservedRunningTime="2026-03-17 15:41:58.995818118 +0000 UTC m=+310.409134165" Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.152589 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562702-hxck5"] Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.153913 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562702-hxck5" Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.156513 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.156662 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.156864 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.169547 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562702-hxck5"] Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.177674 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzk8m\" (UniqueName: \"kubernetes.io/projected/6670565d-1552-47bf-83c5-bf2cadda7f03-kube-api-access-zzk8m\") pod \"auto-csr-approver-29562702-hxck5\" (UID: \"6670565d-1552-47bf-83c5-bf2cadda7f03\") " pod="openshift-infra/auto-csr-approver-29562702-hxck5" Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.278783 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzk8m\" (UniqueName: \"kubernetes.io/projected/6670565d-1552-47bf-83c5-bf2cadda7f03-kube-api-access-zzk8m\") pod \"auto-csr-approver-29562702-hxck5\" (UID: \"6670565d-1552-47bf-83c5-bf2cadda7f03\") " pod="openshift-infra/auto-csr-approver-29562702-hxck5" Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.301090 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzk8m\" (UniqueName: \"kubernetes.io/projected/6670565d-1552-47bf-83c5-bf2cadda7f03-kube-api-access-zzk8m\") pod \"auto-csr-approver-29562702-hxck5\" (UID: \"6670565d-1552-47bf-83c5-bf2cadda7f03\") " pod="openshift-infra/auto-csr-approver-29562702-hxck5" Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.521465 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562702-hxck5" Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.947443 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562702-hxck5"] Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.973377 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2kvv" event={"ID":"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a","Type":"ContainerStarted","Data":"d712664e0621936c7e42f48e9a7ff3550206cf644bd6afb4d6cd443650560dd7"} Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.978627 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562702-hxck5" event={"ID":"6670565d-1552-47bf-83c5-bf2cadda7f03","Type":"ContainerStarted","Data":"d814df982a24ec304fe48ed9312d986bca734ce3b6d0e735358dd5e9a10fcd55"} Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.980781 4767 generic.go:334] "Generic (PLEG): container finished" podID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerID="e76f0e7f83dec65bda69231c082bc4ce8cccc7fa902c2933d856c707983c139f" exitCode=0 Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.980863 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhw26" event={"ID":"ebb17b29-f33c-4236-bb9d-034821b17ba3","Type":"ContainerDied","Data":"e76f0e7f83dec65bda69231c082bc4ce8cccc7fa902c2933d856c707983c139f"} Mar 17 15:42:00 crc kubenswrapper[4767]: I0317 15:42:00.986859 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq699" event={"ID":"7d1d71c5-161c-4ba3-96da-ed07128ff6e1","Type":"ContainerStarted","Data":"ec1c2377f115e0370d4144ab2eecc434c4819d536ad1233e16c156dd49d8c13b"} Mar 17 15:42:01 crc kubenswrapper[4767]: I0317 15:42:01.006231 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d2kvv" podStartSLOduration=8.613270669 podStartE2EDuration="1m41.006215623s" podCreationTimestamp="2026-03-17 15:40:20 +0000 UTC" firstStartedPulling="2026-03-17 15:40:27.373921431 +0000 UTC m=+218.787237478" lastFinishedPulling="2026-03-17 15:41:59.766866385 +0000 UTC m=+311.180182432" observedRunningTime="2026-03-17 15:42:01.003212025 +0000 UTC m=+312.416528082" watchObservedRunningTime="2026-03-17 15:42:01.006215623 +0000 UTC m=+312.419531660" Mar 17 15:42:01 crc kubenswrapper[4767]: I0317 15:42:01.051734 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rq699" podStartSLOduration=6.657092132 podStartE2EDuration="1m40.051713597s" podCreationTimestamp="2026-03-17 15:40:21 +0000 UTC" firstStartedPulling="2026-03-17 15:40:27.309219022 +0000 UTC m=+218.722535069" lastFinishedPulling="2026-03-17 15:42:00.703840487 +0000 UTC m=+312.117156534" observedRunningTime="2026-03-17 15:42:01.045908127 +0000 UTC m=+312.459224194" watchObservedRunningTime="2026-03-17 15:42:01.051713597 +0000 UTC m=+312.465029644" Mar 17 15:42:01 crc kubenswrapper[4767]: I0317 15:42:01.546612 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8m5cc"] Mar 17 15:42:01 crc kubenswrapper[4767]: I0317 15:42:01.740411 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:42:01 crc kubenswrapper[4767]: I0317 15:42:01.742048 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:42:01 crc kubenswrapper[4767]: I0317 15:42:01.930798 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:42:02 crc kubenswrapper[4767]: I0317 15:42:02.053257 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:42:02 crc kubenswrapper[4767]: I0317 15:42:02.085697 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rq699" Mar 17 15:42:02 crc kubenswrapper[4767]: I0317 15:42:02.085770 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rq699" Mar 17 15:42:02 crc kubenswrapper[4767]: I0317 15:42:02.999108 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhw26" event={"ID":"ebb17b29-f33c-4236-bb9d-034821b17ba3","Type":"ContainerStarted","Data":"47d3e38c707c00ec4e13fb95f34e2fe86dcd452b34bdf6602b64eaebd64ed9e1"} Mar 17 15:42:03 crc kubenswrapper[4767]: I0317 15:42:03.138077 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rq699" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerName="registry-server" probeResult="failure" output=< Mar 17 15:42:03 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 15:42:03 crc kubenswrapper[4767]: > Mar 17 15:42:04 crc kubenswrapper[4767]: I0317 15:42:04.008780 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562702-hxck5" event={"ID":"6670565d-1552-47bf-83c5-bf2cadda7f03","Type":"ContainerStarted","Data":"dd01e68ebb275da7b94cc346d8698cbae46ab1a1d65664f44c2feea325983ccf"} Mar 17 15:42:04 crc kubenswrapper[4767]: I0317 15:42:04.030290 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562702-hxck5" podStartSLOduration=1.8740687999999999 podStartE2EDuration="4.030251311s" podCreationTimestamp="2026-03-17 15:42:00 +0000 UTC" firstStartedPulling="2026-03-17 15:42:00.965229482 +0000 UTC m=+312.378545529" lastFinishedPulling="2026-03-17 15:42:03.121411983 +0000 UTC m=+314.534728040" observedRunningTime="2026-03-17 15:42:04.026528642 +0000 UTC m=+315.439844709" watchObservedRunningTime="2026-03-17 15:42:04.030251311 +0000 UTC m=+315.443567358" Mar 17 15:42:04 crc kubenswrapper[4767]: I0317 15:42:04.048348 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vhw26" podStartSLOduration=7.831591413 podStartE2EDuration="1m41.048316611s" podCreationTimestamp="2026-03-17 15:40:23 +0000 UTC" firstStartedPulling="2026-03-17 15:40:28.786839548 +0000 UTC m=+220.200155595" lastFinishedPulling="2026-03-17 15:42:02.003564746 +0000 UTC m=+313.416880793" observedRunningTime="2026-03-17 15:42:04.046832927 +0000 UTC m=+315.460148994" watchObservedRunningTime="2026-03-17 15:42:04.048316611 +0000 UTC m=+315.461632658" Mar 17 15:42:04 crc kubenswrapper[4767]: I0317 15:42:04.540080 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:42:04 crc kubenswrapper[4767]: I0317 15:42:04.541091 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:42:04 crc kubenswrapper[4767]: I0317 15:42:04.614961 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:42:05 crc kubenswrapper[4767]: I0317 15:42:05.018615 4767 generic.go:334] "Generic (PLEG): container finished" podID="6670565d-1552-47bf-83c5-bf2cadda7f03" containerID="dd01e68ebb275da7b94cc346d8698cbae46ab1a1d65664f44c2feea325983ccf" exitCode=0 Mar 17 15:42:05 crc kubenswrapper[4767]: I0317 15:42:05.018696 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562702-hxck5" event={"ID":"6670565d-1552-47bf-83c5-bf2cadda7f03","Type":"ContainerDied","Data":"dd01e68ebb275da7b94cc346d8698cbae46ab1a1d65664f44c2feea325983ccf"} Mar 17 15:42:05 crc kubenswrapper[4767]: I0317 15:42:05.100828 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:42:05 crc kubenswrapper[4767]: I0317 15:42:05.146976 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:42:05 crc kubenswrapper[4767]: I0317 15:42:05.147067 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:42:05 crc kubenswrapper[4767]: I0317 15:42:05.206548 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:42:05 crc kubenswrapper[4767]: I0317 15:42:05.206630 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:42:05 crc kubenswrapper[4767]: I0317 15:42:05.219664 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:42:05 crc kubenswrapper[4767]: I0317 15:42:05.374753 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:42:05 crc kubenswrapper[4767]: I0317 15:42:05.425489 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:42:06 crc kubenswrapper[4767]: I0317 15:42:06.087408 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:42:06 crc kubenswrapper[4767]: I0317 15:42:06.259548 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vhw26" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerName="registry-server" probeResult="failure" output=< Mar 17 15:42:06 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 15:42:06 crc kubenswrapper[4767]: > Mar 17 15:42:06 crc kubenswrapper[4767]: I0317 15:42:06.348571 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562702-hxck5" Mar 17 15:42:06 crc kubenswrapper[4767]: I0317 15:42:06.473116 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzk8m\" (UniqueName: \"kubernetes.io/projected/6670565d-1552-47bf-83c5-bf2cadda7f03-kube-api-access-zzk8m\") pod \"6670565d-1552-47bf-83c5-bf2cadda7f03\" (UID: \"6670565d-1552-47bf-83c5-bf2cadda7f03\") " Mar 17 15:42:06 crc kubenswrapper[4767]: I0317 15:42:06.479457 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6670565d-1552-47bf-83c5-bf2cadda7f03-kube-api-access-zzk8m" (OuterVolumeSpecName: "kube-api-access-zzk8m") pod "6670565d-1552-47bf-83c5-bf2cadda7f03" (UID: "6670565d-1552-47bf-83c5-bf2cadda7f03"). InnerVolumeSpecName "kube-api-access-zzk8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:42:06 crc kubenswrapper[4767]: I0317 15:42:06.574518 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzk8m\" (UniqueName: \"kubernetes.io/projected/6670565d-1552-47bf-83c5-bf2cadda7f03-kube-api-access-zzk8m\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:07 crc kubenswrapper[4767]: I0317 15:42:07.035808 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562702-hxck5" event={"ID":"6670565d-1552-47bf-83c5-bf2cadda7f03","Type":"ContainerDied","Data":"d814df982a24ec304fe48ed9312d986bca734ce3b6d0e735358dd5e9a10fcd55"} Mar 17 15:42:07 crc kubenswrapper[4767]: I0317 15:42:07.035863 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d814df982a24ec304fe48ed9312d986bca734ce3b6d0e735358dd5e9a10fcd55" Mar 17 15:42:07 crc kubenswrapper[4767]: I0317 15:42:07.035936 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562702-hxck5" Mar 17 15:42:07 crc kubenswrapper[4767]: I0317 15:42:07.451551 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxsb9"] Mar 17 15:42:08 crc kubenswrapper[4767]: I0317 15:42:08.041471 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zxsb9" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" containerName="registry-server" containerID="cri-o://0a1bddc789e7b1c9b052b88d48562c8043c6d450b5ff48b838b7c57ed7d25024" gracePeriod=2 Mar 17 15:42:08 crc kubenswrapper[4767]: I0317 15:42:08.643861 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b96464484-gwtw5"] Mar 17 15:42:08 crc kubenswrapper[4767]: I0317 15:42:08.645622 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" containerName="controller-manager" containerID="cri-o://f5bc9e45c59aa87b62bc954daf73e543ba9fb8e158d07ceffe6d501b83a0c856" gracePeriod=30 Mar 17 15:42:08 crc kubenswrapper[4767]: I0317 15:42:08.743014 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw"] Mar 17 15:42:08 crc kubenswrapper[4767]: I0317 15:42:08.743341 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" containerName="route-controller-manager" containerID="cri-o://aea7085a2c9a5f6e02a18c90d1bc6249ac3cc93e159b69e1c4ef5bbab41925e6" gracePeriod=30 Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.106254 4767 generic.go:334] "Generic (PLEG): container finished" podID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" containerID="aea7085a2c9a5f6e02a18c90d1bc6249ac3cc93e159b69e1c4ef5bbab41925e6" exitCode=0 Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.106615 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" event={"ID":"06e548a4-cb50-4864-ba89-e33e8a3dda4e","Type":"ContainerDied","Data":"aea7085a2c9a5f6e02a18c90d1bc6249ac3cc93e159b69e1c4ef5bbab41925e6"} Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.111317 4767 generic.go:334] "Generic (PLEG): container finished" podID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" containerID="f5bc9e45c59aa87b62bc954daf73e543ba9fb8e158d07ceffe6d501b83a0c856" exitCode=0 Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.111419 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" event={"ID":"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b","Type":"ContainerDied","Data":"f5bc9e45c59aa87b62bc954daf73e543ba9fb8e158d07ceffe6d501b83a0c856"} Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.131006 4767 generic.go:334] "Generic (PLEG): container finished" podID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" containerID="0a1bddc789e7b1c9b052b88d48562c8043c6d450b5ff48b838b7c57ed7d25024" exitCode=0 Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.131065 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxsb9" event={"ID":"3f4f4f7f-0983-4551-9a1c-88c6214322f4","Type":"ContainerDied","Data":"0a1bddc789e7b1c9b052b88d48562c8043c6d450b5ff48b838b7c57ed7d25024"} Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.160448 4767 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.160857 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6670565d-1552-47bf-83c5-bf2cadda7f03" containerName="oc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.160869 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6670565d-1552-47bf-83c5-bf2cadda7f03" containerName="oc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.160975 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="6670565d-1552-47bf-83c5-bf2cadda7f03" containerName="oc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.161432 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.162390 4767 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.162935 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f" gracePeriod=15 Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.163032 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2" gracePeriod=15 Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.163123 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff" gracePeriod=15 Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.163228 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0" gracePeriod=15 Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.163309 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29" gracePeriod=15 Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164078 4767 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.164452 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164474 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.164488 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164496 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.164510 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164529 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.164539 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164546 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.164560 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164568 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.164584 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164590 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.164603 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164610 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.164626 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164633 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164757 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164767 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164777 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164788 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164794 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164801 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164811 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.164917 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164925 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.164934 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.164940 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.165092 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.165105 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.234198 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.235950 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.236070 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.236148 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.236216 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.236251 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.236310 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.236332 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.236377 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337580 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337632 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337697 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337728 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337744 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337767 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337770 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337735 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337836 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337852 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337781 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337872 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337906 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337828 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337866 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.337990 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.359728 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.360027 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.434225 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.434887 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.435246 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.437140 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.437782 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.438140 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.438537 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.535885 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.539348 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-config\") pod \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.539493 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e548a4-cb50-4864-ba89-e33e8a3dda4e-serving-cert\") pod \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.539522 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jfsc\" (UniqueName: \"kubernetes.io/projected/06e548a4-cb50-4864-ba89-e33e8a3dda4e-kube-api-access-5jfsc\") pod \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.539590 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhfdf\" (UniqueName: \"kubernetes.io/projected/3f4f4f7f-0983-4551-9a1c-88c6214322f4-kube-api-access-xhfdf\") pod \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.539611 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-client-ca\") pod \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\" (UID: \"06e548a4-cb50-4864-ba89-e33e8a3dda4e\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.539795 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-utilities\") pod \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.539825 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-catalog-content\") pod \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\" (UID: \"3f4f4f7f-0983-4551-9a1c-88c6214322f4\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.540548 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-client-ca" (OuterVolumeSpecName: "client-ca") pod "06e548a4-cb50-4864-ba89-e33e8a3dda4e" (UID: "06e548a4-cb50-4864-ba89-e33e8a3dda4e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.540578 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-config" (OuterVolumeSpecName: "config") pod "06e548a4-cb50-4864-ba89-e33e8a3dda4e" (UID: "06e548a4-cb50-4864-ba89-e33e8a3dda4e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.541138 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-utilities" (OuterVolumeSpecName: "utilities") pod "3f4f4f7f-0983-4551-9a1c-88c6214322f4" (UID: "3f4f4f7f-0983-4551-9a1c-88c6214322f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.548451 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e548a4-cb50-4864-ba89-e33e8a3dda4e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "06e548a4-cb50-4864-ba89-e33e8a3dda4e" (UID: "06e548a4-cb50-4864-ba89-e33e8a3dda4e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.565638 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f4f4f7f-0983-4551-9a1c-88c6214322f4-kube-api-access-xhfdf" (OuterVolumeSpecName: "kube-api-access-xhfdf") pod "3f4f4f7f-0983-4551-9a1c-88c6214322f4" (UID: "3f4f4f7f-0983-4551-9a1c-88c6214322f4"). InnerVolumeSpecName "kube-api-access-xhfdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.568848 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f4f4f7f-0983-4551-9a1c-88c6214322f4" (UID: "3f4f4f7f-0983-4551-9a1c-88c6214322f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.570489 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06e548a4-cb50-4864-ba89-e33e8a3dda4e-kube-api-access-5jfsc" (OuterVolumeSpecName: "kube-api-access-5jfsc") pod "06e548a4-cb50-4864-ba89-e33e8a3dda4e" (UID: "06e548a4-cb50-4864-ba89-e33e8a3dda4e"). InnerVolumeSpecName "kube-api-access-5jfsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: E0317 15:42:09.576456 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.119:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189dab4251676a1f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:42:09.575766559 +0000 UTC m=+320.989082596,LastTimestamp:2026-03-17 15:42:09.575766559 +0000 UTC m=+320.989082596,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.641057 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e548a4-cb50-4864-ba89-e33e8a3dda4e-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.641124 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jfsc\" (UniqueName: \"kubernetes.io/projected/06e548a4-cb50-4864-ba89-e33e8a3dda4e-kube-api-access-5jfsc\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.641140 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhfdf\" (UniqueName: \"kubernetes.io/projected/3f4f4f7f-0983-4551-9a1c-88c6214322f4-kube-api-access-xhfdf\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.641152 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.641191 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.641208 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4f4f7f-0983-4551-9a1c-88c6214322f4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.641223 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e548a4-cb50-4864-ba89-e33e8a3dda4e-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.776235 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.776991 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.777472 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.777681 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.778196 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.843423 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-config\") pod \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.843578 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-serving-cert\") pod \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.843612 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g9tx\" (UniqueName: \"kubernetes.io/projected/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-kube-api-access-2g9tx\") pod \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.843656 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-proxy-ca-bundles\") pod \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.843690 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-client-ca\") pod \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\" (UID: \"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b\") " Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.844951 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-client-ca" (OuterVolumeSpecName: "client-ca") pod "e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" (UID: "e9bbbb6f-72aa-4d0d-809b-566c9f176d6b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.844900 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" (UID: "e9bbbb6f-72aa-4d0d-809b-566c9f176d6b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.845024 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-config" (OuterVolumeSpecName: "config") pod "e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" (UID: "e9bbbb6f-72aa-4d0d-809b-566c9f176d6b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.849145 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-kube-api-access-2g9tx" (OuterVolumeSpecName: "kube-api-access-2g9tx") pod "e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" (UID: "e9bbbb6f-72aa-4d0d-809b-566c9f176d6b"). InnerVolumeSpecName "kube-api-access-2g9tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.849140 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" (UID: "e9bbbb6f-72aa-4d0d-809b-566c9f176d6b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.945116 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.945149 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.945158 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.945179 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g9tx\" (UniqueName: \"kubernetes.io/projected/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-kube-api-access-2g9tx\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:09 crc kubenswrapper[4767]: I0317 15:42:09.945190 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.139212 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" event={"ID":"e9bbbb6f-72aa-4d0d-809b-566c9f176d6b","Type":"ContainerDied","Data":"1ac61668b742c00028227095883917a88f4e8ad912716050d8e27710ab880554"} Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.139231 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.139293 4767 scope.go:117] "RemoveContainer" containerID="f5bc9e45c59aa87b62bc954daf73e543ba9fb8e158d07ceffe6d501b83a0c856" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.141711 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.142363 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.142397 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.142777 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.143517 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.144113 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.145027 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2" exitCode=0 Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.145086 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff" exitCode=0 Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.145098 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0" exitCode=0 Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.145108 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29" exitCode=2 Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.147566 4767 generic.go:334] "Generic (PLEG): container finished" podID="d49401a4-1f38-4e26-92d4-69c2085d6640" containerID="ea084c55031cdaf7e3942e6d2ab50b9eeb84bdb7ec2bd283c3a8e06c0d8953f0" exitCode=0 Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.147615 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d49401a4-1f38-4e26-92d4-69c2085d6640","Type":"ContainerDied","Data":"ea084c55031cdaf7e3942e6d2ab50b9eeb84bdb7ec2bd283c3a8e06c0d8953f0"} Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.154339 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.154816 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.155562 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.155895 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.156360 4767 scope.go:117] "RemoveContainer" containerID="8726bf3459d9af17a050986495e86947412c7a098155a214c3216eb84e0a08f8" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.156455 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.162786 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.163059 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.163412 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.163616 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.163835 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.173072 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxsb9" event={"ID":"3f4f4f7f-0983-4551-9a1c-88c6214322f4","Type":"ContainerDied","Data":"b01bb10f584c109c735f4dc8778b8716940f893878b44dfd96eabee8babfdb90"} Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.173255 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxsb9" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.175531 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.175902 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.176117 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.176643 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.177060 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.178748 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" event={"ID":"06e548a4-cb50-4864-ba89-e33e8a3dda4e","Type":"ContainerDied","Data":"640bd0bd4c9604176ebf62f8e39d92793b2859c6056c7bcc42388a6e73e7e6bb"} Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.178798 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.179731 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.179929 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.180111 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.180332 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.180502 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.182542 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"2dbc28d9ffd954d2d73d861230550162a96fddc3fbddf7a86ea70372a0aa6344"} Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.183111 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"e2f4049fe82794df48e5036af235ed3696bb83448c85e50ee068a60b03070199"} Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.183924 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.184241 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.184506 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.184908 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.185843 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.189795 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.190085 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.190402 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.190766 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.191008 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.193890 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.194220 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.194533 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.194765 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.194982 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.196068 4767 scope.go:117] "RemoveContainer" containerID="0a1bddc789e7b1c9b052b88d48562c8043c6d450b5ff48b838b7c57ed7d25024" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.214889 4767 scope.go:117] "RemoveContainer" containerID="83cf83ceebcb7c8c1ed7e69e335c1a5d086be3dc6d45ff7f0b4889b99c79e99c" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.235909 4767 scope.go:117] "RemoveContainer" containerID="15c6634662eb24ce7a354c47d11199fe27a893e241b63dc425e5977b70a2a0ed" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.251449 4767 scope.go:117] "RemoveContainer" containerID="aea7085a2c9a5f6e02a18c90d1bc6249ac3cc93e159b69e1c4ef5bbab41925e6" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.795763 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.795813 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.854829 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.855588 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.856288 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.856711 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.857271 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.857701 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:10 crc kubenswrapper[4767]: I0317 15:42:10.858031 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.195560 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.286536 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.289278 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.289546 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.289753 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.290894 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.291229 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.291571 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.556242 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.557140 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.557530 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.557960 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.558383 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.558585 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.558854 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.563514 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.564236 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.564990 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.565470 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.565746 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.566043 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.566348 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.566586 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.566865 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.680387 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-kubelet-dir\") pod \"d49401a4-1f38-4e26-92d4-69c2085d6640\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.680438 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-var-lock\") pod \"d49401a4-1f38-4e26-92d4-69c2085d6640\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.680482 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d49401a4-1f38-4e26-92d4-69c2085d6640" (UID: "d49401a4-1f38-4e26-92d4-69c2085d6640"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.680574 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-var-lock" (OuterVolumeSpecName: "var-lock") pod "d49401a4-1f38-4e26-92d4-69c2085d6640" (UID: "d49401a4-1f38-4e26-92d4-69c2085d6640"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.680611 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d49401a4-1f38-4e26-92d4-69c2085d6640-kube-api-access\") pod \"d49401a4-1f38-4e26-92d4-69c2085d6640\" (UID: \"d49401a4-1f38-4e26-92d4-69c2085d6640\") " Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.680629 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.680657 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.680682 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.680699 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.680734 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.680790 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.681047 4767 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.681069 4767 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.681080 4767 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.681091 4767 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.681101 4767 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d49401a4-1f38-4e26-92d4-69c2085d6640-var-lock\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.686460 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d49401a4-1f38-4e26-92d4-69c2085d6640-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d49401a4-1f38-4e26-92d4-69c2085d6640" (UID: "d49401a4-1f38-4e26-92d4-69c2085d6640"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:42:11 crc kubenswrapper[4767]: I0317 15:42:11.782562 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d49401a4-1f38-4e26-92d4-69c2085d6640-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.123183 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rq699" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.123704 4767 status_manager.go:851] "Failed to get status for pod" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" pod="openshift-marketplace/community-operators-rq699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rq699\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.124039 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.124684 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.125446 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.125870 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.126605 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.126934 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.127248 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.170016 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rq699" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.170420 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.170760 4767 status_manager.go:851] "Failed to get status for pod" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" pod="openshift-marketplace/community-operators-rq699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rq699\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.170998 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.171246 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.171562 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.171823 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.172044 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.172369 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.209124 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.210401 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f" exitCode=0 Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.210511 4767 scope.go:117] "RemoveContainer" containerID="0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.210547 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.213667 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d49401a4-1f38-4e26-92d4-69c2085d6640","Type":"ContainerDied","Data":"303663cdb320caf53b07662e4a191abf5704bfc4ceaa09559c86a55da22d9218"} Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.213733 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="303663cdb320caf53b07662e4a191abf5704bfc4ceaa09559c86a55da22d9218" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.213864 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.228414 4767 scope.go:117] "RemoveContainer" containerID="07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.229319 4767 status_manager.go:851] "Failed to get status for pod" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" pod="openshift-marketplace/community-operators-rq699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rq699\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.229616 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.230082 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.230670 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.231545 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.232291 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.232613 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.232844 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.234032 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.234407 4767 status_manager.go:851] "Failed to get status for pod" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" pod="openshift-marketplace/community-operators-rq699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rq699\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.234748 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.234983 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.235255 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.235541 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.235880 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.236105 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.242965 4767 scope.go:117] "RemoveContainer" containerID="c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.255675 4767 scope.go:117] "RemoveContainer" containerID="c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.270515 4767 scope.go:117] "RemoveContainer" containerID="06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.289529 4767 scope.go:117] "RemoveContainer" containerID="e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.306935 4767 scope.go:117] "RemoveContainer" containerID="0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2" Mar 17 15:42:12 crc kubenswrapper[4767]: E0317 15:42:12.307429 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\": container with ID starting with 0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2 not found: ID does not exist" containerID="0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.307472 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2"} err="failed to get container status \"0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\": rpc error: code = NotFound desc = could not find container \"0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2\": container with ID starting with 0d3c4c2b0c151f5c8cdccb7a392e58d0ff65f94f92227f84e5b9f6665aca04f2 not found: ID does not exist" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.307503 4767 scope.go:117] "RemoveContainer" containerID="07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff" Mar 17 15:42:12 crc kubenswrapper[4767]: E0317 15:42:12.307983 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\": container with ID starting with 07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff not found: ID does not exist" containerID="07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.308011 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff"} err="failed to get container status \"07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\": rpc error: code = NotFound desc = could not find container \"07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff\": container with ID starting with 07faa5ca06729cb2d1d731e801bc7c7917590554500f4a85810ce9864abcbdff not found: ID does not exist" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.308029 4767 scope.go:117] "RemoveContainer" containerID="c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0" Mar 17 15:42:12 crc kubenswrapper[4767]: E0317 15:42:12.308283 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\": container with ID starting with c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0 not found: ID does not exist" containerID="c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.308323 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0"} err="failed to get container status \"c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\": rpc error: code = NotFound desc = could not find container \"c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0\": container with ID starting with c085c1270d4dce1eb2920eccb709c0dde590c278c26416910b9b58b1232241a0 not found: ID does not exist" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.308341 4767 scope.go:117] "RemoveContainer" containerID="c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29" Mar 17 15:42:12 crc kubenswrapper[4767]: E0317 15:42:12.308623 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\": container with ID starting with c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29 not found: ID does not exist" containerID="c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.308643 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29"} err="failed to get container status \"c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\": rpc error: code = NotFound desc = could not find container \"c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29\": container with ID starting with c3c8287454cc2d373e2d8eeeacef62310796cae79a90f77a98976c1b7e606f29 not found: ID does not exist" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.308655 4767 scope.go:117] "RemoveContainer" containerID="06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f" Mar 17 15:42:12 crc kubenswrapper[4767]: E0317 15:42:12.308938 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\": container with ID starting with 06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f not found: ID does not exist" containerID="06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.308959 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f"} err="failed to get container status \"06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\": rpc error: code = NotFound desc = could not find container \"06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f\": container with ID starting with 06856332794da19c65e2f17f25c7545a368a99db3a1ff2799f62c1638a99529f not found: ID does not exist" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.308972 4767 scope.go:117] "RemoveContainer" containerID="e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f" Mar 17 15:42:12 crc kubenswrapper[4767]: E0317 15:42:12.309323 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\": container with ID starting with e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f not found: ID does not exist" containerID="e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f" Mar 17 15:42:12 crc kubenswrapper[4767]: I0317 15:42:12.309361 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f"} err="failed to get container status \"e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\": rpc error: code = NotFound desc = could not find container \"e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f\": container with ID starting with e86bf727c071062ebda9bd6527281d20937a10ebd671b5b39021158f668f2c7f not found: ID does not exist" Mar 17 15:42:13 crc kubenswrapper[4767]: I0317 15:42:13.403098 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.251683 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.252676 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.252945 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.253402 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.253735 4767 status_manager.go:851] "Failed to get status for pod" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" pod="openshift-marketplace/redhat-operators-vhw26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vhw26\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.253981 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.254231 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.254682 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.254924 4767 status_manager.go:851] "Failed to get status for pod" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" pod="openshift-marketplace/community-operators-rq699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rq699\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.286331 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.286977 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.287405 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.287927 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.288514 4767 status_manager.go:851] "Failed to get status for pod" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" pod="openshift-marketplace/community-operators-rq699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rq699\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.288808 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.289061 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.289381 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:15 crc kubenswrapper[4767]: I0317 15:42:15.289620 4767 status_manager.go:851] "Failed to get status for pod" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" pod="openshift-marketplace/redhat-operators-vhw26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vhw26\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:18 crc kubenswrapper[4767]: E0317 15:42:18.421359 4767 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:18 crc kubenswrapper[4767]: E0317 15:42:18.422745 4767 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:18 crc kubenswrapper[4767]: E0317 15:42:18.423416 4767 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:18 crc kubenswrapper[4767]: E0317 15:42:18.423682 4767 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:18 crc kubenswrapper[4767]: E0317 15:42:18.423882 4767 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:18 crc kubenswrapper[4767]: I0317 15:42:18.423930 4767 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 17 15:42:18 crc kubenswrapper[4767]: E0317 15:42:18.424109 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" interval="200ms" Mar 17 15:42:18 crc kubenswrapper[4767]: E0317 15:42:18.625525 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" interval="400ms" Mar 17 15:42:18 crc kubenswrapper[4767]: E0317 15:42:18.772186 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.119:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189dab4251676a1f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 15:42:09.575766559 +0000 UTC m=+320.989082596,LastTimestamp:2026-03-17 15:42:09.575766559 +0000 UTC m=+320.989082596,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 15:42:19 crc kubenswrapper[4767]: E0317 15:42:19.026996 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" interval="800ms" Mar 17 15:42:19 crc kubenswrapper[4767]: I0317 15:42:19.357466 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:19 crc kubenswrapper[4767]: I0317 15:42:19.358162 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:19 crc kubenswrapper[4767]: I0317 15:42:19.358409 4767 status_manager.go:851] "Failed to get status for pod" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" pod="openshift-marketplace/community-operators-rq699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rq699\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:19 crc kubenswrapper[4767]: I0317 15:42:19.358670 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:19 crc kubenswrapper[4767]: I0317 15:42:19.358881 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:19 crc kubenswrapper[4767]: I0317 15:42:19.359088 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:19 crc kubenswrapper[4767]: I0317 15:42:19.359371 4767 status_manager.go:851] "Failed to get status for pod" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" pod="openshift-marketplace/redhat-operators-vhw26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vhw26\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:19 crc kubenswrapper[4767]: I0317 15:42:19.359687 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:19 crc kubenswrapper[4767]: E0317 15:42:19.828589 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" interval="1.6s" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.354139 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.355521 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.356161 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.356544 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.357042 4767 status_manager.go:851] "Failed to get status for pod" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" pod="openshift-marketplace/redhat-operators-vhw26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vhw26\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.357388 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.357661 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.357872 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.358033 4767 status_manager.go:851] "Failed to get status for pod" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" pod="openshift-marketplace/community-operators-rq699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rq699\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.370947 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.370973 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:20 crc kubenswrapper[4767]: E0317 15:42:20.371255 4767 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.371813 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:20 crc kubenswrapper[4767]: W0317 15:42:20.402679 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-a9296051db196d4ddf5eebab9e78d54e76458af06c72adebf8dcd571fbfa6c22 WatchSource:0}: Error finding container a9296051db196d4ddf5eebab9e78d54e76458af06c72adebf8dcd571fbfa6c22: Status 404 returned error can't find the container with id a9296051db196d4ddf5eebab9e78d54e76458af06c72adebf8dcd571fbfa6c22 Mar 17 15:42:20 crc kubenswrapper[4767]: I0317 15:42:20.445481 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a9296051db196d4ddf5eebab9e78d54e76458af06c72adebf8dcd571fbfa6c22"} Mar 17 15:42:21 crc kubenswrapper[4767]: E0317 15:42:21.430323 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.119:6443: connect: connection refused" interval="3.2s" Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.451595 4767 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="3ded1cd29b4e9ecba088d3f4f56416864a91a50ef27a3c9fa477e4d386646936" exitCode=0 Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.451707 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"3ded1cd29b4e9ecba088d3f4f56416864a91a50ef27a3c9fa477e4d386646936"} Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.451921 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.451945 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:21 crc kubenswrapper[4767]: E0317 15:42:21.452411 4767 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.452473 4767 status_manager.go:851] "Failed to get status for pod" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.452954 4767 status_manager.go:851] "Failed to get status for pod" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" pod="openshift-controller-manager/controller-manager-6b96464484-gwtw5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6b96464484-gwtw5\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.453209 4767 status_manager.go:851] "Failed to get status for pod" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" pod="openshift-marketplace/redhat-operators-vhw26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vhw26\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.453483 4767 status_manager.go:851] "Failed to get status for pod" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" pod="openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7697ffcff9-hbkkw\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.453737 4767 status_manager.go:851] "Failed to get status for pod" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" pod="openshift-marketplace/certified-operators-d2kvv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-d2kvv\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.453952 4767 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.454241 4767 status_manager.go:851] "Failed to get status for pod" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" pod="openshift-marketplace/community-operators-rq699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rq699\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:21 crc kubenswrapper[4767]: I0317 15:42:21.454690 4767 status_manager.go:851] "Failed to get status for pod" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" pod="openshift-marketplace/redhat-marketplace-zxsb9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zxsb9\": dial tcp 38.102.83.119:6443: connect: connection refused" Mar 17 15:42:22 crc kubenswrapper[4767]: I0317 15:42:22.467516 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"78066c66b5ec9faf62cc9c1d186ffe03f1674900956afb8052872dbdb43f18cb"} Mar 17 15:42:22 crc kubenswrapper[4767]: I0317 15:42:22.468018 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"678b529e460316e0aa35d340f9a3d357e3b272308f1ef53e7d96010e520379b0"} Mar 17 15:42:22 crc kubenswrapper[4767]: I0317 15:42:22.468032 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4d5469d167d2e3ce5cfab6f7253fa7ddef326eaa10db7d01f6681f7c3c0fde87"} Mar 17 15:42:22 crc kubenswrapper[4767]: I0317 15:42:22.468043 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0a892deb47c4a9822d17a3e1db94f507480c01de00a23db0b2be2ac9c84264c7"} Mar 17 15:42:23 crc kubenswrapper[4767]: I0317 15:42:23.476815 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7b735a438057ad6120ddf42c919981195afe8c9eff0331cc9578b26bae544af9"} Mar 17 15:42:23 crc kubenswrapper[4767]: I0317 15:42:23.477460 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:23 crc kubenswrapper[4767]: I0317 15:42:23.477726 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:23 crc kubenswrapper[4767]: I0317 15:42:23.477760 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:24 crc kubenswrapper[4767]: I0317 15:42:24.486074 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 17 15:42:24 crc kubenswrapper[4767]: I0317 15:42:24.490450 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 17 15:42:24 crc kubenswrapper[4767]: I0317 15:42:24.490543 4767 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="1a6d14de833771c3a060a2879802f4f7f37b96b60004b4e5f96374e5fee1428e" exitCode=1 Mar 17 15:42:24 crc kubenswrapper[4767]: I0317 15:42:24.490596 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"1a6d14de833771c3a060a2879802f4f7f37b96b60004b4e5f96374e5fee1428e"} Mar 17 15:42:24 crc kubenswrapper[4767]: I0317 15:42:24.491400 4767 scope.go:117] "RemoveContainer" containerID="1a6d14de833771c3a060a2879802f4f7f37b96b60004b4e5f96374e5fee1428e" Mar 17 15:42:25 crc kubenswrapper[4767]: I0317 15:42:25.372340 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:25 crc kubenswrapper[4767]: I0317 15:42:25.372643 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:25 crc kubenswrapper[4767]: I0317 15:42:25.377432 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:25 crc kubenswrapper[4767]: I0317 15:42:25.500240 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 17 15:42:25 crc kubenswrapper[4767]: I0317 15:42:25.502890 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 17 15:42:25 crc kubenswrapper[4767]: I0317 15:42:25.502945 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1a0bfdf7badf045b41c62ef25a6c37928c0abf036ce202175f2dd3cbc88a647c"} Mar 17 15:42:26 crc kubenswrapper[4767]: I0317 15:42:26.573396 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" podUID="0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" containerName="oauth-openshift" containerID="cri-o://c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4" gracePeriod=15 Mar 17 15:42:26 crc kubenswrapper[4767]: I0317 15:42:26.963868 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.260610 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-login\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.261319 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-cliconfig\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.261525 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mj6f\" (UniqueName: \"kubernetes.io/projected/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-kube-api-access-7mj6f\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.261640 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-error\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.261758 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-router-certs\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.261960 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-dir\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.262091 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-ocp-branding-template\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.261952 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.261991 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.262214 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-serving-cert\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.262525 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-service-ca\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.262645 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-idp-0-file-data\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.262738 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-policies\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.262905 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-provider-selection\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.263243 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.263283 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.263571 4767 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.263681 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.263786 4767 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.263866 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.270732 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.272627 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.272764 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-kube-api-access-7mj6f" (OuterVolumeSpecName: "kube-api-access-7mj6f") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "kube-api-access-7mj6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.274958 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.280407 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.282543 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.289301 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.291617 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.364307 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-trusted-ca-bundle\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.364361 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-session\") pod \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\" (UID: \"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1\") " Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.364531 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mj6f\" (UniqueName: \"kubernetes.io/projected/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-kube-api-access-7mj6f\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.364542 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.364562 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.364573 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.364582 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.364591 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.364601 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.364610 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.364882 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.367131 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" (UID: "0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.465498 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.465726 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.520370 4767 generic.go:334] "Generic (PLEG): container finished" podID="0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" containerID="c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4" exitCode=0 Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.520406 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" event={"ID":"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1","Type":"ContainerDied","Data":"c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4"} Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.520429 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" event={"ID":"0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1","Type":"ContainerDied","Data":"8aa4d12dd0d870e2f48e6ecfadf1a14cac81bc479e952be42fc44b83d06db29d"} Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.520448 4767 scope.go:117] "RemoveContainer" containerID="c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.520471 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8m5cc" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.541417 4767 scope.go:117] "RemoveContainer" containerID="c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4" Mar 17 15:42:27 crc kubenswrapper[4767]: E0317 15:42:27.542196 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4\": container with ID starting with c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4 not found: ID does not exist" containerID="c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4" Mar 17 15:42:27 crc kubenswrapper[4767]: I0317 15:42:27.542246 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4"} err="failed to get container status \"c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4\": rpc error: code = NotFound desc = could not find container \"c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4\": container with ID starting with c2b38f286f4ece537e3500811c68f976b43d47cd451736912fc4a0b5101678e4 not found: ID does not exist" Mar 17 15:42:28 crc kubenswrapper[4767]: I0317 15:42:28.585938 4767 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:29 crc kubenswrapper[4767]: I0317 15:42:29.375777 4767 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="2cad33eb-44bd-40fd-8530-e82833aa18b7" Mar 17 15:42:29 crc kubenswrapper[4767]: E0317 15:42:29.479800 4767 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"audit\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Mar 17 15:42:29 crc kubenswrapper[4767]: I0317 15:42:29.532365 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:29 crc kubenswrapper[4767]: I0317 15:42:29.532604 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:29 crc kubenswrapper[4767]: I0317 15:42:29.536093 4767 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="2cad33eb-44bd-40fd-8530-e82833aa18b7" Mar 17 15:42:29 crc kubenswrapper[4767]: I0317 15:42:29.537355 4767 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://0a892deb47c4a9822d17a3e1db94f507480c01de00a23db0b2be2ac9c84264c7" Mar 17 15:42:29 crc kubenswrapper[4767]: I0317 15:42:29.537386 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:29 crc kubenswrapper[4767]: I0317 15:42:29.975282 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:42:30 crc kubenswrapper[4767]: I0317 15:42:30.538220 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:30 crc kubenswrapper[4767]: I0317 15:42:30.538632 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:30 crc kubenswrapper[4767]: I0317 15:42:30.540891 4767 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="2cad33eb-44bd-40fd-8530-e82833aa18b7" Mar 17 15:42:32 crc kubenswrapper[4767]: I0317 15:42:32.330996 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:42:32 crc kubenswrapper[4767]: I0317 15:42:32.334628 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:42:38 crc kubenswrapper[4767]: I0317 15:42:38.273929 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 17 15:42:38 crc kubenswrapper[4767]: I0317 15:42:38.406009 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 17 15:42:38 crc kubenswrapper[4767]: I0317 15:42:38.961533 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 17 15:42:39 crc kubenswrapper[4767]: I0317 15:42:39.835371 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 17 15:42:39 crc kubenswrapper[4767]: I0317 15:42:39.981690 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 15:42:40 crc kubenswrapper[4767]: I0317 15:42:40.957946 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 17 15:42:40 crc kubenswrapper[4767]: I0317 15:42:40.980450 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 17 15:42:41 crc kubenswrapper[4767]: I0317 15:42:41.277690 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 17 15:42:41 crc kubenswrapper[4767]: I0317 15:42:41.449860 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 17 15:42:41 crc kubenswrapper[4767]: I0317 15:42:41.862270 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 17 15:42:41 crc kubenswrapper[4767]: I0317 15:42:41.946673 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.373336 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.400890 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.500102 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.566387 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.619439 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.626280 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.752897 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.908786 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.909052 4767 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.909331 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.909662 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 17 15:42:42 crc kubenswrapper[4767]: I0317 15:42:42.910027 4767 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.009309 4767 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.204611 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.243284 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.384261 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.414024 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.483312 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.513802 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.625641 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.672964 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.676065 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.849253 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.963706 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.973264 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 17 15:42:43 crc kubenswrapper[4767]: I0317 15:42:43.975642 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.099571 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.132355 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.266777 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.348849 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.446803 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.537534 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.727891 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.788282 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.812602 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.820821 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.827972 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.889664 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 17 15:42:44 crc kubenswrapper[4767]: I0317 15:42:44.958316 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.048006 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.050389 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.063925 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.081454 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.165775 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.166368 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.221603 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.229235 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.554072 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.554197 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.596984 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.647907 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.650220 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.661506 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.664745 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.720108 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.799455 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.814252 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.821893 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.922557 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.958852 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.976547 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 17 15:42:45 crc kubenswrapper[4767]: I0317 15:42:45.995701 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.018869 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.112739 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.166637 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.228752 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.275825 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.562108 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.570615 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.570927 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.786536 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.890770 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.897472 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.897696 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.908675 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.909088 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.915791 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 17 15:42:46 crc kubenswrapper[4767]: I0317 15:42:46.924658 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.128803 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.159535 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.205034 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.252123 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.272646 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.273863 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.317078 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.323230 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.351611 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.478031 4767 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.493467 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.504126 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.506258 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.593271 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.603714 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.612413 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.642155 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.779379 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.878868 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.880738 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.888523 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 17 15:42:47 crc kubenswrapper[4767]: I0317 15:42:47.902804 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.007905 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.081050 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.081392 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.209586 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.236314 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.365745 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.367028 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.392944 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.537140 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.549122 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.571687 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.951590 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.951816 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.960243 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.960429 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.960677 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.960958 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 17 15:42:48 crc kubenswrapper[4767]: I0317 15:42:48.961129 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.004821 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.004848 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.087412 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.184303 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.191964 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.269250 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.312105 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.338106 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.339230 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.586064 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.610470 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.644411 4767 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.647410 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=40.647389874 podStartE2EDuration="40.647389874s" podCreationTimestamp="2026-03-17 15:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:42:28.683717849 +0000 UTC m=+340.097033906" watchObservedRunningTime="2026-03-17 15:42:49.647389874 +0000 UTC m=+361.060705921" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.650965 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6b96464484-gwtw5","openshift-marketplace/redhat-marketplace-zxsb9","openshift-authentication/oauth-openshift-558db77b4-8m5cc","openshift-route-controller-manager/route-controller-manager-7697ffcff9-hbkkw","openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651031 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5cc5b65bd-6nll7","openshift-controller-manager/controller-manager-57bc79d78-25d96","openshift-kube-apiserver/kube-apiserver-crc","openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8"] Mar 17 15:42:49 crc kubenswrapper[4767]: E0317 15:42:49.651248 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" containerName="oauth-openshift" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651262 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" containerName="oauth-openshift" Mar 17 15:42:49 crc kubenswrapper[4767]: E0317 15:42:49.651272 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" containerName="extract-utilities" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651279 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" containerName="extract-utilities" Mar 17 15:42:49 crc kubenswrapper[4767]: E0317 15:42:49.651286 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" containerName="controller-manager" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651294 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" containerName="controller-manager" Mar 17 15:42:49 crc kubenswrapper[4767]: E0317 15:42:49.651304 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" containerName="installer" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651310 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" containerName="installer" Mar 17 15:42:49 crc kubenswrapper[4767]: E0317 15:42:49.651318 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" containerName="registry-server" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651323 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" containerName="registry-server" Mar 17 15:42:49 crc kubenswrapper[4767]: E0317 15:42:49.651337 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" containerName="route-controller-manager" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651343 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" containerName="route-controller-manager" Mar 17 15:42:49 crc kubenswrapper[4767]: E0317 15:42:49.651352 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" containerName="extract-content" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651357 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" containerName="extract-content" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651444 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" containerName="route-controller-manager" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651454 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" containerName="registry-server" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651463 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d49401a4-1f38-4e26-92d4-69c2085d6640" containerName="installer" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651473 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" containerName="controller-manager" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.651481 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" containerName="oauth-openshift" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.652213 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.652319 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.652972 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.653089 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="786771d9-2ad9-49e9-9c2f-f86e587539a8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.653487 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.654905 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.655842 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.655961 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.656760 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.656998 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.657009 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.657271 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.657512 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.657553 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.657649 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.657766 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.657995 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658224 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658402 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1767bf72-c95d-486d-baa6-ca741742569b-config\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658445 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-template-login\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658490 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5v27\" (UniqueName: \"kubernetes.io/projected/2321f9fb-78c8-401b-8fc2-a5f82d42a226-kube-api-access-l5v27\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658516 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-service-ca\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658569 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1767bf72-c95d-486d-baa6-ca741742569b-proxy-ca-bundles\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658586 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658594 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658600 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658658 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658683 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-audit-policies\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658716 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-session\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658745 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1767bf72-c95d-486d-baa6-ca741742569b-client-ca\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658771 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-router-certs\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658820 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9f5b566-a215-4322-8d74-668dc2a93f3e-config\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658861 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658895 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9znh\" (UniqueName: \"kubernetes.io/projected/a9f5b566-a215-4322-8d74-668dc2a93f3e-kube-api-access-x9znh\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658908 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658929 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx5cv\" (UniqueName: \"kubernetes.io/projected/1767bf72-c95d-486d-baa6-ca741742569b-kube-api-access-wx5cv\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658937 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658953 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.658978 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.659052 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9f5b566-a215-4322-8d74-668dc2a93f3e-client-ca\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.659081 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f5b566-a215-4322-8d74-668dc2a93f3e-serving-cert\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.659122 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.659108 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.659211 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1767bf72-c95d-486d-baa6-ca741742569b-serving-cert\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.659231 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-template-error\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.659256 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2321f9fb-78c8-401b-8fc2-a5f82d42a226-audit-dir\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.659977 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.660267 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.660380 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.660540 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.661498 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.667912 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.670165 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.672339 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.677458 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.679356 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.719550 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.719534483 podStartE2EDuration="21.719534483s" podCreationTimestamp="2026-03-17 15:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:42:49.719297635 +0000 UTC m=+361.132613702" watchObservedRunningTime="2026-03-17 15:42:49.719534483 +0000 UTC m=+361.132850530" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.759106 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.759642 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9f5b566-a215-4322-8d74-668dc2a93f3e-client-ca\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.759688 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f5b566-a215-4322-8d74-668dc2a93f3e-serving-cert\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.759717 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.759747 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1767bf72-c95d-486d-baa6-ca741742569b-serving-cert\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.759769 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-template-error\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.759793 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2321f9fb-78c8-401b-8fc2-a5f82d42a226-audit-dir\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760234 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2321f9fb-78c8-401b-8fc2-a5f82d42a226-audit-dir\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760579 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1767bf72-c95d-486d-baa6-ca741742569b-config\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760620 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-template-login\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760644 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5v27\" (UniqueName: \"kubernetes.io/projected/2321f9fb-78c8-401b-8fc2-a5f82d42a226-kube-api-access-l5v27\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760673 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-service-ca\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760695 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1767bf72-c95d-486d-baa6-ca741742569b-proxy-ca-bundles\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760768 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760806 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760836 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-audit-policies\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760861 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-session\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760890 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1767bf72-c95d-486d-baa6-ca741742569b-client-ca\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760917 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-router-certs\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760944 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9f5b566-a215-4322-8d74-668dc2a93f3e-config\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760952 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9f5b566-a215-4322-8d74-668dc2a93f3e-client-ca\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760974 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.760996 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9znh\" (UniqueName: \"kubernetes.io/projected/a9f5b566-a215-4322-8d74-668dc2a93f3e-kube-api-access-x9znh\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.761076 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx5cv\" (UniqueName: \"kubernetes.io/projected/1767bf72-c95d-486d-baa6-ca741742569b-kube-api-access-wx5cv\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.761102 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.761123 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.761339 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.761968 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1767bf72-c95d-486d-baa6-ca741742569b-config\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.762029 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.762318 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-service-ca\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.762404 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9f5b566-a215-4322-8d74-668dc2a93f3e-config\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.763477 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2321f9fb-78c8-401b-8fc2-a5f82d42a226-audit-policies\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.764440 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1767bf72-c95d-486d-baa6-ca741742569b-client-ca\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.764672 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1767bf72-c95d-486d-baa6-ca741742569b-proxy-ca-bundles\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.765028 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-template-error\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.765047 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f5b566-a215-4322-8d74-668dc2a93f3e-serving-cert\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.765158 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-template-login\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.765542 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-session\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.766268 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.766601 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-router-certs\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.766979 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.767632 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.768112 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2321f9fb-78c8-401b-8fc2-a5f82d42a226-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.768679 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1767bf72-c95d-486d-baa6-ca741742569b-serving-cert\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.779810 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5v27\" (UniqueName: \"kubernetes.io/projected/2321f9fb-78c8-401b-8fc2-a5f82d42a226-kube-api-access-l5v27\") pod \"oauth-openshift-5cc5b65bd-6nll7\" (UID: \"2321f9fb-78c8-401b-8fc2-a5f82d42a226\") " pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.782564 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9znh\" (UniqueName: \"kubernetes.io/projected/a9f5b566-a215-4322-8d74-668dc2a93f3e-kube-api-access-x9znh\") pod \"route-controller-manager-56d7bf8c9b-cs8b8\" (UID: \"a9f5b566-a215-4322-8d74-668dc2a93f3e\") " pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.782746 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx5cv\" (UniqueName: \"kubernetes.io/projected/1767bf72-c95d-486d-baa6-ca741742569b-kube-api-access-wx5cv\") pod \"controller-manager-57bc79d78-25d96\" (UID: \"1767bf72-c95d-486d-baa6-ca741742569b\") " pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.895148 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.954080 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.968736 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.974711 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.985308 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:49 crc kubenswrapper[4767]: I0317 15:42:49.992901 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.148921 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.229213 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.254753 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.480238 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.491578 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.501603 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.514628 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.519898 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8"] Mar 17 15:42:50 crc kubenswrapper[4767]: W0317 15:42:50.526120 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9f5b566_a215_4322_8d74_668dc2a93f3e.slice/crio-7e1219fd6ed5322cf8dee929fd48c795f43db605c934bb590acbaa050128fa32 WatchSource:0}: Error finding container 7e1219fd6ed5322cf8dee929fd48c795f43db605c934bb590acbaa050128fa32: Status 404 returned error can't find the container with id 7e1219fd6ed5322cf8dee929fd48c795f43db605c934bb590acbaa050128fa32 Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.570397 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.594490 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.645607 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.683682 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.684378 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.689400 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.694342 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.704470 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.737025 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5cc5b65bd-6nll7"] Mar 17 15:42:50 crc kubenswrapper[4767]: W0317 15:42:50.744385 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2321f9fb_78c8_401b_8fc2_a5f82d42a226.slice/crio-afdb89871db8772850e91e2a54783b331e9d8b4bc37d0d99c7f693acb0b1a611 WatchSource:0}: Error finding container afdb89871db8772850e91e2a54783b331e9d8b4bc37d0d99c7f693acb0b1a611: Status 404 returned error can't find the container with id afdb89871db8772850e91e2a54783b331e9d8b4bc37d0d99c7f693acb0b1a611 Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.755615 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.763735 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57bc79d78-25d96"] Mar 17 15:42:50 crc kubenswrapper[4767]: W0317 15:42:50.770868 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1767bf72_c95d_486d_baa6_ca741742569b.slice/crio-8ab38c562b46aa99f81695e4bd7dadbcff3ed24d2e4c9eceb0d0cf5a909901c6 WatchSource:0}: Error finding container 8ab38c562b46aa99f81695e4bd7dadbcff3ed24d2e4c9eceb0d0cf5a909901c6: Status 404 returned error can't find the container with id 8ab38c562b46aa99f81695e4bd7dadbcff3ed24d2e4c9eceb0d0cf5a909901c6 Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.786654 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.908093 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.969937 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" event={"ID":"2321f9fb-78c8-401b-8fc2-a5f82d42a226","Type":"ContainerStarted","Data":"afdb89871db8772850e91e2a54783b331e9d8b4bc37d0d99c7f693acb0b1a611"} Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.971196 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" event={"ID":"1767bf72-c95d-486d-baa6-ca741742569b","Type":"ContainerStarted","Data":"07a6477a1b291c54a3879e589f04d6266ea243a0214e6d194b7b8b54cd46e3f2"} Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.971228 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" event={"ID":"1767bf72-c95d-486d-baa6-ca741742569b","Type":"ContainerStarted","Data":"8ab38c562b46aa99f81695e4bd7dadbcff3ed24d2e4c9eceb0d0cf5a909901c6"} Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.972005 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.973655 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" event={"ID":"a9f5b566-a215-4322-8d74-668dc2a93f3e","Type":"ContainerStarted","Data":"7568ac5efdde80256b6a21b3c20a6b083df449905aa94a3be7af802805f2636f"} Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.973681 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.973690 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" event={"ID":"a9f5b566-a215-4322-8d74-668dc2a93f3e","Type":"ContainerStarted","Data":"7e1219fd6ed5322cf8dee929fd48c795f43db605c934bb590acbaa050128fa32"} Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.974650 4767 patch_prober.go:28] interesting pod/controller-manager-57bc79d78-25d96 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Mar 17 15:42:50 crc kubenswrapper[4767]: I0317 15:42:50.974685 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podUID="1767bf72-c95d-486d-baa6-ca741742569b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.015078 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podStartSLOduration=43.015064073 podStartE2EDuration="43.015064073s" podCreationTimestamp="2026-03-17 15:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:42:50.998543571 +0000 UTC m=+362.411859628" watchObservedRunningTime="2026-03-17 15:42:51.015064073 +0000 UTC m=+362.428380120" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.016507 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" podStartSLOduration=43.016500708 podStartE2EDuration="43.016500708s" podCreationTimestamp="2026-03-17 15:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:42:51.012499182 +0000 UTC m=+362.425815229" watchObservedRunningTime="2026-03-17 15:42:51.016500708 +0000 UTC m=+362.429816755" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.146233 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.236637 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.247319 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.307727 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.313221 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.351219 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.360932 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06e548a4-cb50-4864-ba89-e33e8a3dda4e" path="/var/lib/kubelet/pods/06e548a4-cb50-4864-ba89-e33e8a3dda4e/volumes" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.361662 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1" path="/var/lib/kubelet/pods/0771bb7a-0ab0-43be-ade4-6f6a4f1ebbd1/volumes" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.362412 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f4f4f7f-0983-4551-9a1c-88c6214322f4" path="/var/lib/kubelet/pods/3f4f4f7f-0983-4551-9a1c-88c6214322f4/volumes" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.363875 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9bbbb6f-72aa-4d0d-809b-566c9f176d6b" path="/var/lib/kubelet/pods/e9bbbb6f-72aa-4d0d-809b-566c9f176d6b/volumes" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.385519 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.394540 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.749098 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.749340 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.749116 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.749647 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.754672 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.782624 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.786146 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.873795 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.993020 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" event={"ID":"2321f9fb-78c8-401b-8fc2-a5f82d42a226","Type":"ContainerStarted","Data":"eaff0e00c90c96aca126c5718dbe75b4480b9cf9712683c7700af0e35030c21d"} Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.993506 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:51 crc kubenswrapper[4767]: I0317 15:42:51.998633 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 15:42:52 crc kubenswrapper[4767]: I0317 15:42:52.007096 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 17 15:42:52 crc kubenswrapper[4767]: I0317 15:42:52.016828 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 17 15:42:52 crc kubenswrapper[4767]: I0317 15:42:52.019145 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" podStartSLOduration=51.019129867 podStartE2EDuration="51.019129867s" podCreationTimestamp="2026-03-17 15:42:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:42:52.017343981 +0000 UTC m=+363.430660048" watchObservedRunningTime="2026-03-17 15:42:52.019129867 +0000 UTC m=+363.432445914" Mar 17 15:42:52 crc kubenswrapper[4767]: I0317 15:42:52.076303 4767 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 17 15:42:52 crc kubenswrapper[4767]: I0317 15:42:52.141859 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 17 15:42:52 crc kubenswrapper[4767]: I0317 15:42:52.194162 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 17 15:42:52 crc kubenswrapper[4767]: I0317 15:42:52.236806 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 15:42:52 crc kubenswrapper[4767]: I0317 15:42:52.579797 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 17 15:42:52 crc kubenswrapper[4767]: I0317 15:42:52.865241 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 17 15:42:52 crc kubenswrapper[4767]: I0317 15:42:52.898395 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 17 15:42:52 crc kubenswrapper[4767]: I0317 15:42:52.978844 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.031803 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.037642 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.116530 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.187248 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.234014 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.306766 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.325825 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.404035 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.570240 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.584865 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.714157 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.722364 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 17 15:42:53 crc kubenswrapper[4767]: I0317 15:42:53.744883 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 17 15:42:54 crc kubenswrapper[4767]: I0317 15:42:54.121593 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 17 15:42:54 crc kubenswrapper[4767]: I0317 15:42:54.122226 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 17 15:42:54 crc kubenswrapper[4767]: I0317 15:42:54.380990 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 17 15:42:54 crc kubenswrapper[4767]: I0317 15:42:54.404849 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 17 15:42:54 crc kubenswrapper[4767]: I0317 15:42:54.475910 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 17 15:42:55 crc kubenswrapper[4767]: I0317 15:42:55.131510 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 17 15:42:55 crc kubenswrapper[4767]: I0317 15:42:55.133315 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 17 15:42:55 crc kubenswrapper[4767]: I0317 15:42:55.133486 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 17 15:42:55 crc kubenswrapper[4767]: I0317 15:42:55.133659 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 17 15:42:55 crc kubenswrapper[4767]: I0317 15:42:55.162337 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 17 15:42:55 crc kubenswrapper[4767]: I0317 15:42:55.172787 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 17 15:42:55 crc kubenswrapper[4767]: I0317 15:42:55.246822 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 17 15:42:55 crc kubenswrapper[4767]: I0317 15:42:55.368006 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 17 15:42:55 crc kubenswrapper[4767]: I0317 15:42:55.495763 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 17 15:42:55 crc kubenswrapper[4767]: I0317 15:42:55.865650 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 17 15:42:56 crc kubenswrapper[4767]: I0317 15:42:56.008362 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 17 15:42:56 crc kubenswrapper[4767]: I0317 15:42:56.607779 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 17 15:42:56 crc kubenswrapper[4767]: I0317 15:42:56.733812 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 17 15:42:57 crc kubenswrapper[4767]: I0317 15:42:57.013545 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 17 15:42:57 crc kubenswrapper[4767]: I0317 15:42:57.037711 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 17 15:42:57 crc kubenswrapper[4767]: I0317 15:42:57.040360 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 17 15:42:57 crc kubenswrapper[4767]: I0317 15:42:57.117393 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 17 15:42:57 crc kubenswrapper[4767]: I0317 15:42:57.218343 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 17 15:42:57 crc kubenswrapper[4767]: I0317 15:42:57.334870 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 17 15:42:57 crc kubenswrapper[4767]: I0317 15:42:57.462334 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 17 15:42:58 crc kubenswrapper[4767]: I0317 15:42:58.087275 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 17 15:43:01 crc kubenswrapper[4767]: I0317 15:43:01.386205 4767 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 17 15:43:01 crc kubenswrapper[4767]: I0317 15:43:01.386732 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://2dbc28d9ffd954d2d73d861230550162a96fddc3fbddf7a86ea70372a0aa6344" gracePeriod=5 Mar 17 15:43:06 crc kubenswrapper[4767]: I0317 15:43:06.823163 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 17 15:43:06 crc kubenswrapper[4767]: I0317 15:43:06.824203 4767 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="2dbc28d9ffd954d2d73d861230550162a96fddc3fbddf7a86ea70372a0aa6344" exitCode=137 Mar 17 15:43:06 crc kubenswrapper[4767]: I0317 15:43:06.968403 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 17 15:43:06 crc kubenswrapper[4767]: I0317 15:43:06.968479 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.100270 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.100372 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.100503 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.100532 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.100619 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.100619 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.100662 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.100698 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.100828 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.101134 4767 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.101149 4767 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.101204 4767 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.101217 4767 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.112596 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.201929 4767 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.449043 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.449390 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.467697 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.467742 4767 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="c31b33ee-f80e-468e-bda2-b909c328019e" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.473466 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.473569 4767 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="c31b33ee-f80e-468e-bda2-b909c328019e" Mar 17 15:43:07 crc kubenswrapper[4767]: E0317 15:43:07.554903 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-e2f4049fe82794df48e5036af235ed3696bb83448c85e50ee068a60b03070199\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice\": RecentStats: unable to find data in memory cache]" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.832579 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.832716 4767 scope.go:117] "RemoveContainer" containerID="2dbc28d9ffd954d2d73d861230550162a96fddc3fbddf7a86ea70372a0aa6344" Mar 17 15:43:07 crc kubenswrapper[4767]: I0317 15:43:07.832773 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.593092 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d2kvv"] Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.594036 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d2kvv" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" containerName="registry-server" containerID="cri-o://d712664e0621936c7e42f48e9a7ff3550206cf644bd6afb4d6cd443650560dd7" gracePeriod=30 Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.606153 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-klkzj"] Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.606452 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-klkzj" podUID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" containerName="registry-server" containerID="cri-o://b2b986e0261578449e1f42241f3d405db374fb6e7e52966e8a4003ce254adbf3" gracePeriod=30 Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.624715 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rq699"] Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.625136 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rq699" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerName="registry-server" containerID="cri-o://ec1c2377f115e0370d4144ab2eecc434c4819d536ad1233e16c156dd49d8c13b" gracePeriod=30 Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.630908 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cb6sk"] Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.631264 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" podUID="acdeabaa-85fb-4d85-bad6-89fc1be0125a" containerName="marketplace-operator" containerID="cri-o://eae33d5c864c011827785ea5feaf0c9e88f81a1c3da2b35d6871e80fab77a840" gracePeriod=30 Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.635424 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7ktd"] Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.635627 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c7ktd" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerName="registry-server" containerID="cri-o://81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208" gracePeriod=30 Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.655030 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q2n5r"] Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.655914 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q2n5r" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerName="registry-server" containerID="cri-o://e90359a0dfe7155779e0ad33fa338d37c63ebe2059107a4279d83d66ee56158b" gracePeriod=30 Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.979922 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vl8vb"] Mar 17 15:43:12 crc kubenswrapper[4767]: E0317 15:43:12.981358 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.981500 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.981876 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.983810 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vhw26"] Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.984351 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vhw26" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerName="registry-server" containerID="cri-o://47d3e38c707c00ec4e13fb95f34e2fe86dcd452b34bdf6602b64eaebd64ed9e1" gracePeriod=30 Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.984887 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:12 crc kubenswrapper[4767]: I0317 15:43:12.992533 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vl8vb"] Mar 17 15:43:13 crc kubenswrapper[4767]: I0317 15:43:13.031049 4767 generic.go:334] "Generic (PLEG): container finished" podID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" containerID="d712664e0621936c7e42f48e9a7ff3550206cf644bd6afb4d6cd443650560dd7" exitCode=0 Mar 17 15:43:13 crc kubenswrapper[4767]: I0317 15:43:13.031102 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2kvv" event={"ID":"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a","Type":"ContainerDied","Data":"d712664e0621936c7e42f48e9a7ff3550206cf644bd6afb4d6cd443650560dd7"} Mar 17 15:43:13 crc kubenswrapper[4767]: I0317 15:43:13.073388 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/eafb476c-0d04-4608-b1e3-cf8b6ebb74bf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vl8vb\" (UID: \"eafb476c-0d04-4608-b1e3-cf8b6ebb74bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:13 crc kubenswrapper[4767]: I0317 15:43:13.073450 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmcmv\" (UniqueName: \"kubernetes.io/projected/eafb476c-0d04-4608-b1e3-cf8b6ebb74bf-kube-api-access-mmcmv\") pod \"marketplace-operator-79b997595-vl8vb\" (UID: \"eafb476c-0d04-4608-b1e3-cf8b6ebb74bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:13 crc kubenswrapper[4767]: I0317 15:43:13.073474 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eafb476c-0d04-4608-b1e3-cf8b6ebb74bf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vl8vb\" (UID: \"eafb476c-0d04-4608-b1e3-cf8b6ebb74bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:13 crc kubenswrapper[4767]: I0317 15:43:13.174951 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/eafb476c-0d04-4608-b1e3-cf8b6ebb74bf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vl8vb\" (UID: \"eafb476c-0d04-4608-b1e3-cf8b6ebb74bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:13 crc kubenswrapper[4767]: I0317 15:43:13.175105 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmcmv\" (UniqueName: \"kubernetes.io/projected/eafb476c-0d04-4608-b1e3-cf8b6ebb74bf-kube-api-access-mmcmv\") pod \"marketplace-operator-79b997595-vl8vb\" (UID: \"eafb476c-0d04-4608-b1e3-cf8b6ebb74bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:13 crc kubenswrapper[4767]: I0317 15:43:13.175151 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eafb476c-0d04-4608-b1e3-cf8b6ebb74bf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vl8vb\" (UID: \"eafb476c-0d04-4608-b1e3-cf8b6ebb74bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:13 crc kubenswrapper[4767]: I0317 15:43:13.179594 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eafb476c-0d04-4608-b1e3-cf8b6ebb74bf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vl8vb\" (UID: \"eafb476c-0d04-4608-b1e3-cf8b6ebb74bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:13 crc kubenswrapper[4767]: I0317 15:43:13.183809 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/eafb476c-0d04-4608-b1e3-cf8b6ebb74bf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vl8vb\" (UID: \"eafb476c-0d04-4608-b1e3-cf8b6ebb74bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:13 crc kubenswrapper[4767]: I0317 15:43:13.196268 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmcmv\" (UniqueName: \"kubernetes.io/projected/eafb476c-0d04-4608-b1e3-cf8b6ebb74bf-kube-api-access-mmcmv\") pod \"marketplace-operator-79b997595-vl8vb\" (UID: \"eafb476c-0d04-4608-b1e3-cf8b6ebb74bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.101396 4767 generic.go:334] "Generic (PLEG): container finished" podID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerID="47d3e38c707c00ec4e13fb95f34e2fe86dcd452b34bdf6602b64eaebd64ed9e1" exitCode=0 Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.101652 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhw26" event={"ID":"ebb17b29-f33c-4236-bb9d-034821b17ba3","Type":"ContainerDied","Data":"47d3e38c707c00ec4e13fb95f34e2fe86dcd452b34bdf6602b64eaebd64ed9e1"} Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.110306 4767 generic.go:334] "Generic (PLEG): container finished" podID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerID="81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208" exitCode=0 Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.110381 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7ktd" event={"ID":"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3","Type":"ContainerDied","Data":"81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208"} Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.112380 4767 generic.go:334] "Generic (PLEG): container finished" podID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerID="ec1c2377f115e0370d4144ab2eecc434c4819d536ad1233e16c156dd49d8c13b" exitCode=0 Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.112424 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq699" event={"ID":"7d1d71c5-161c-4ba3-96da-ed07128ff6e1","Type":"ContainerDied","Data":"ec1c2377f115e0370d4144ab2eecc434c4819d536ad1233e16c156dd49d8c13b"} Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.119512 4767 generic.go:334] "Generic (PLEG): container finished" podID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" containerID="b2b986e0261578449e1f42241f3d405db374fb6e7e52966e8a4003ce254adbf3" exitCode=0 Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.119601 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkzj" event={"ID":"3dc430ad-a22a-4fe8-a4dd-5c94622da680","Type":"ContainerDied","Data":"b2b986e0261578449e1f42241f3d405db374fb6e7e52966e8a4003ce254adbf3"} Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.125847 4767 generic.go:334] "Generic (PLEG): container finished" podID="acdeabaa-85fb-4d85-bad6-89fc1be0125a" containerID="eae33d5c864c011827785ea5feaf0c9e88f81a1c3da2b35d6871e80fab77a840" exitCode=0 Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.125940 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" event={"ID":"acdeabaa-85fb-4d85-bad6-89fc1be0125a","Type":"ContainerDied","Data":"eae33d5c864c011827785ea5feaf0c9e88f81a1c3da2b35d6871e80fab77a840"} Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.129954 4767 generic.go:334] "Generic (PLEG): container finished" podID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerID="e90359a0dfe7155779e0ad33fa338d37c63ebe2059107a4279d83d66ee56158b" exitCode=0 Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.129999 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2n5r" event={"ID":"e76337aa-429a-49b9-99d3-e261c32ee85b","Type":"ContainerDied","Data":"e90359a0dfe7155779e0ad33fa338d37c63ebe2059107a4279d83d66ee56158b"} Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.195004 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.329568 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.496217 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9x85\" (UniqueName: \"kubernetes.io/projected/acdeabaa-85fb-4d85-bad6-89fc1be0125a-kube-api-access-g9x85\") pod \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.496739 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-operator-metrics\") pod \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.496777 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-trusted-ca\") pod \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\" (UID: \"acdeabaa-85fb-4d85-bad6-89fc1be0125a\") " Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.498123 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "acdeabaa-85fb-4d85-bad6-89fc1be0125a" (UID: "acdeabaa-85fb-4d85-bad6-89fc1be0125a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.511862 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "acdeabaa-85fb-4d85-bad6-89fc1be0125a" (UID: "acdeabaa-85fb-4d85-bad6-89fc1be0125a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.512670 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acdeabaa-85fb-4d85-bad6-89fc1be0125a-kube-api-access-g9x85" (OuterVolumeSpecName: "kube-api-access-g9x85") pod "acdeabaa-85fb-4d85-bad6-89fc1be0125a" (UID: "acdeabaa-85fb-4d85-bad6-89fc1be0125a"). InnerVolumeSpecName "kube-api-access-g9x85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:43:14 crc kubenswrapper[4767]: E0317 15:43:14.540502 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208 is running failed: container process not found" containerID="81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 15:43:14 crc kubenswrapper[4767]: E0317 15:43:14.540911 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208 is running failed: container process not found" containerID="81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 15:43:14 crc kubenswrapper[4767]: E0317 15:43:14.541439 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208 is running failed: container process not found" containerID="81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 15:43:14 crc kubenswrapper[4767]: E0317 15:43:14.541479 4767 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-c7ktd" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerName="registry-server" Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.599509 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9x85\" (UniqueName: \"kubernetes.io/projected/acdeabaa-85fb-4d85-bad6-89fc1be0125a-kube-api-access-g9x85\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.599551 4767 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.599565 4767 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/acdeabaa-85fb-4d85-bad6-89fc1be0125a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.656567 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:43:14 crc kubenswrapper[4767]: I0317 15:43:14.670468 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:14.802924 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-catalog-content\") pod \"e76337aa-429a-49b9-99d3-e261c32ee85b\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:14.803068 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-catalog-content\") pod \"ebb17b29-f33c-4236-bb9d-034821b17ba3\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:14.803149 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-utilities\") pod \"ebb17b29-f33c-4236-bb9d-034821b17ba3\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:14.803304 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc7fv\" (UniqueName: \"kubernetes.io/projected/e76337aa-429a-49b9-99d3-e261c32ee85b-kube-api-access-xc7fv\") pod \"e76337aa-429a-49b9-99d3-e261c32ee85b\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:14.803399 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdbm5\" (UniqueName: \"kubernetes.io/projected/ebb17b29-f33c-4236-bb9d-034821b17ba3-kube-api-access-cdbm5\") pod \"ebb17b29-f33c-4236-bb9d-034821b17ba3\" (UID: \"ebb17b29-f33c-4236-bb9d-034821b17ba3\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:14.803445 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-utilities\") pod \"e76337aa-429a-49b9-99d3-e261c32ee85b\" (UID: \"e76337aa-429a-49b9-99d3-e261c32ee85b\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.126526 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-utilities" (OuterVolumeSpecName: "utilities") pod "ebb17b29-f33c-4236-bb9d-034821b17ba3" (UID: "ebb17b29-f33c-4236-bb9d-034821b17ba3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.132726 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-utilities" (OuterVolumeSpecName: "utilities") pod "e76337aa-429a-49b9-99d3-e261c32ee85b" (UID: "e76337aa-429a-49b9-99d3-e261c32ee85b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.137514 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.137556 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.150616 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebb17b29-f33c-4236-bb9d-034821b17ba3-kube-api-access-cdbm5" (OuterVolumeSpecName: "kube-api-access-cdbm5") pod "ebb17b29-f33c-4236-bb9d-034821b17ba3" (UID: "ebb17b29-f33c-4236-bb9d-034821b17ba3"). InnerVolumeSpecName "kube-api-access-cdbm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.167888 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e76337aa-429a-49b9-99d3-e261c32ee85b-kube-api-access-xc7fv" (OuterVolumeSpecName: "kube-api-access-xc7fv") pod "e76337aa-429a-49b9-99d3-e261c32ee85b" (UID: "e76337aa-429a-49b9-99d3-e261c32ee85b"). InnerVolumeSpecName "kube-api-access-xc7fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.186376 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.188924 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.189019 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkzj" event={"ID":"3dc430ad-a22a-4fe8-a4dd-5c94622da680","Type":"ContainerDied","Data":"335a92007e3345a9487740053d8b2f9128b93449780abe80f96ae75127834f06"} Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.189081 4767 scope.go:117] "RemoveContainer" containerID="b2b986e0261578449e1f42241f3d405db374fb6e7e52966e8a4003ce254adbf3" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.190979 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.192075 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cb6sk" event={"ID":"acdeabaa-85fb-4d85-bad6-89fc1be0125a","Type":"ContainerDied","Data":"13fb5834ff825d4c4df1f6eff8c22425de2e947991d4fc85ab45dbfae8310bec"} Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.196000 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.197532 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2kvv" event={"ID":"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a","Type":"ContainerDied","Data":"1c5cba4c2d6d7f1607b59467bc764d17bc2be44476f3445ee25fec9c51f99e3d"} Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.204089 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2n5r" event={"ID":"e76337aa-429a-49b9-99d3-e261c32ee85b","Type":"ContainerDied","Data":"debde7dee9f9c11a29fffe2fda7c1ab4623535bd1b88b83d2e6f1407cec24113"} Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.204283 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2n5r" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.239673 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhw26" event={"ID":"ebb17b29-f33c-4236-bb9d-034821b17ba3","Type":"ContainerDied","Data":"181e06f5e1d81e1cc3b53ceb2a2dec30da8fe62b7322f74b32a24c63d584d7cb"} Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.240772 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhw26" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.243369 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8lw6\" (UniqueName: \"kubernetes.io/projected/3dc430ad-a22a-4fe8-a4dd-5c94622da680-kube-api-access-r8lw6\") pod \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.243484 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-catalog-content\") pod \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.243612 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-catalog-content\") pod \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.243718 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-catalog-content\") pod \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.243806 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-utilities\") pod \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.243976 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lp5b\" (UniqueName: \"kubernetes.io/projected/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-kube-api-access-5lp5b\") pod \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\" (UID: \"7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.244116 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2dfp\" (UniqueName: \"kubernetes.io/projected/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-kube-api-access-q2dfp\") pod \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.244139 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-utilities\") pod \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\" (UID: \"3dc430ad-a22a-4fe8-a4dd-5c94622da680\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.244200 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-utilities\") pod \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\" (UID: \"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.244558 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc7fv\" (UniqueName: \"kubernetes.io/projected/e76337aa-429a-49b9-99d3-e261c32ee85b-kube-api-access-xc7fv\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.244572 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdbm5\" (UniqueName: \"kubernetes.io/projected/ebb17b29-f33c-4236-bb9d-034821b17ba3-kube-api-access-cdbm5\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.247010 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-utilities" (OuterVolumeSpecName: "utilities") pod "f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" (UID: "f46dfeb0-7bd5-4835-a16e-8a117a3db8a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.252525 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-utilities" (OuterVolumeSpecName: "utilities") pod "7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" (UID: "7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.259380 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-kube-api-access-5lp5b" (OuterVolumeSpecName: "kube-api-access-5lp5b") pod "7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" (UID: "7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a"). InnerVolumeSpecName "kube-api-access-5lp5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.261219 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-utilities" (OuterVolumeSpecName: "utilities") pod "3dc430ad-a22a-4fe8-a4dd-5c94622da680" (UID: "3dc430ad-a22a-4fe8-a4dd-5c94622da680"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.265507 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dc430ad-a22a-4fe8-a4dd-5c94622da680-kube-api-access-r8lw6" (OuterVolumeSpecName: "kube-api-access-r8lw6") pod "3dc430ad-a22a-4fe8-a4dd-5c94622da680" (UID: "3dc430ad-a22a-4fe8-a4dd-5c94622da680"). InnerVolumeSpecName "kube-api-access-r8lw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.266622 4767 scope.go:117] "RemoveContainer" containerID="42f59dbaacd0e780dee973ffc979cd875439c6bb96338a5ce64ed135ea0ee59d" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.268162 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7ktd" event={"ID":"f46dfeb0-7bd5-4835-a16e-8a117a3db8a3","Type":"ContainerDied","Data":"fa8520c4808219b785e6d448797823dd736e25bc5358b98a2840588a5ee52388"} Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.268370 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7ktd" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.284489 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e76337aa-429a-49b9-99d3-e261c32ee85b" (UID: "e76337aa-429a-49b9-99d3-e261c32ee85b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.310146 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cb6sk"] Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.320496 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cb6sk"] Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.322026 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-kube-api-access-q2dfp" (OuterVolumeSpecName: "kube-api-access-q2dfp") pod "f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" (UID: "f46dfeb0-7bd5-4835-a16e-8a117a3db8a3"). InnerVolumeSpecName "kube-api-access-q2dfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.322881 4767 scope.go:117] "RemoveContainer" containerID="ac6c43a4f582c64a7a75cd098f8481ec8534a13bc07188100494824bd66afd48" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.325951 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" (UID: "f46dfeb0-7bd5-4835-a16e-8a117a3db8a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.345705 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2dfp\" (UniqueName: \"kubernetes.io/projected/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-kube-api-access-q2dfp\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.345755 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.345770 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.345780 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8lw6\" (UniqueName: \"kubernetes.io/projected/3dc430ad-a22a-4fe8-a4dd-5c94622da680-kube-api-access-r8lw6\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.345790 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.345803 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.345813 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76337aa-429a-49b9-99d3-e261c32ee85b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.345827 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lp5b\" (UniqueName: \"kubernetes.io/projected/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-kube-api-access-5lp5b\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.355104 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3dc430ad-a22a-4fe8-a4dd-5c94622da680" (UID: "3dc430ad-a22a-4fe8-a4dd-5c94622da680"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.360442 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rq699" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.373606 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" (UID: "7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.374026 4767 scope.go:117] "RemoveContainer" containerID="eae33d5c864c011827785ea5feaf0c9e88f81a1c3da2b35d6871e80fab77a840" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.374516 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acdeabaa-85fb-4d85-bad6-89fc1be0125a" path="/var/lib/kubelet/pods/acdeabaa-85fb-4d85-bad6-89fc1be0125a/volumes" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.386141 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebb17b29-f33c-4236-bb9d-034821b17ba3" (UID: "ebb17b29-f33c-4236-bb9d-034821b17ba3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.406798 4767 scope.go:117] "RemoveContainer" containerID="d712664e0621936c7e42f48e9a7ff3550206cf644bd6afb4d6cd443650560dd7" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.426094 4767 scope.go:117] "RemoveContainer" containerID="848997677f41edf0cfe605cd77c42edab072bab43d1fd9876ee6557f9f57c328" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.446735 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-catalog-content\") pod \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.446833 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nm8p\" (UniqueName: \"kubernetes.io/projected/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-kube-api-access-6nm8p\") pod \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.446933 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-utilities\") pod \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\" (UID: \"7d1d71c5-161c-4ba3-96da-ed07128ff6e1\") " Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.447266 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dc430ad-a22a-4fe8-a4dd-5c94622da680-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.447291 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.447301 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb17b29-f33c-4236-bb9d-034821b17ba3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.448282 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-utilities" (OuterVolumeSpecName: "utilities") pod "7d1d71c5-161c-4ba3-96da-ed07128ff6e1" (UID: "7d1d71c5-161c-4ba3-96da-ed07128ff6e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.450765 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-kube-api-access-6nm8p" (OuterVolumeSpecName: "kube-api-access-6nm8p") pod "7d1d71c5-161c-4ba3-96da-ed07128ff6e1" (UID: "7d1d71c5-161c-4ba3-96da-ed07128ff6e1"). InnerVolumeSpecName "kube-api-access-6nm8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.459877 4767 scope.go:117] "RemoveContainer" containerID="3a8ad9a791ce5935f91019adfaf3fbd5123b1ffc65d7e28e0021f28c1ed1b7f5" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.485161 4767 scope.go:117] "RemoveContainer" containerID="e90359a0dfe7155779e0ad33fa338d37c63ebe2059107a4279d83d66ee56158b" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.487634 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vl8vb"] Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.507603 4767 scope.go:117] "RemoveContainer" containerID="8c0e4705b915b89f591cb5140d524eda895c66c18c99ed647317cb1fc9d4d3b3" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.529475 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d1d71c5-161c-4ba3-96da-ed07128ff6e1" (UID: "7d1d71c5-161c-4ba3-96da-ed07128ff6e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.539379 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q2n5r"] Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.545062 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q2n5r"] Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.548150 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nm8p\" (UniqueName: \"kubernetes.io/projected/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-kube-api-access-6nm8p\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.548206 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.548219 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d1d71c5-161c-4ba3-96da-ed07128ff6e1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.550328 4767 scope.go:117] "RemoveContainer" containerID="c75e05a67860ba455fba7d2f51422da56388bdc9c2e87ce454f03509b35ce552" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.568498 4767 scope.go:117] "RemoveContainer" containerID="47d3e38c707c00ec4e13fb95f34e2fe86dcd452b34bdf6602b64eaebd64ed9e1" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.582412 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vhw26"] Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.585312 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vhw26"] Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.608471 4767 scope.go:117] "RemoveContainer" containerID="e76f0e7f83dec65bda69231c082bc4ce8cccc7fa902c2933d856c707983c139f" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.621756 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7ktd"] Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.625787 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7ktd"] Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.657831 4767 scope.go:117] "RemoveContainer" containerID="e3820da4606dbc685eb002165ce2ce61f0b8d5cde93ec9d4290826cde4f4e9fb" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.681712 4767 scope.go:117] "RemoveContainer" containerID="81c529e0ebc07434772680306e92f2264080215c71a219793818f3ca8aace208" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.713551 4767 scope.go:117] "RemoveContainer" containerID="74224c64630e0da55b91e6ac81b9b9b624e553c10f72323f0c1de1748eab9c0a" Mar 17 15:43:15 crc kubenswrapper[4767]: I0317 15:43:15.738508 4767 scope.go:117] "RemoveContainer" containerID="92057d188bd11eb864de3b3a7e87ef9b1e42b9728ae958e64b6665ccb4dda19c" Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.427956 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" event={"ID":"eafb476c-0d04-4608-b1e3-cf8b6ebb74bf","Type":"ContainerStarted","Data":"409b91914e2a0244e2724f3c9fb2419916be4c94b7216e8348717f939bdd1376"} Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.428011 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" event={"ID":"eafb476c-0d04-4608-b1e3-cf8b6ebb74bf","Type":"ContainerStarted","Data":"3b3879519d7affb87b2a1a9ace76daa2ac9b6db992a67bc3714ba7641a3c23f2"} Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.429420 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.433609 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klkzj" Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.437329 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d2kvv" Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.439477 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.444232 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rq699" event={"ID":"7d1d71c5-161c-4ba3-96da-ed07128ff6e1","Type":"ContainerDied","Data":"5eb196e2a1995f63348cd0dbf9c014ae3277f9a4170460732e92d81f4382a2e2"} Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.444292 4767 scope.go:117] "RemoveContainer" containerID="ec1c2377f115e0370d4144ab2eecc434c4819d536ad1233e16c156dd49d8c13b" Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.444395 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rq699" Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.472129 4767 scope.go:117] "RemoveContainer" containerID="17c61359073c5df3cd11758b35ca04cc055d0061fcfd6b7c625bbe56ddc4736a" Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.483991 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" podStartSLOduration=4.4826973500000005 podStartE2EDuration="4.48269735s" podCreationTimestamp="2026-03-17 15:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:43:16.479668504 +0000 UTC m=+387.892984561" watchObservedRunningTime="2026-03-17 15:43:16.48269735 +0000 UTC m=+387.896013397" Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.492580 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-klkzj"] Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.498786 4767 scope.go:117] "RemoveContainer" containerID="1856ee116c91f280b647868400d6c8522c8047bf2bc8ac741f920258bc1e964d" Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.501003 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-klkzj"] Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.536459 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d2kvv"] Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.541543 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d2kvv"] Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.558699 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rq699"] Mar 17 15:43:16 crc kubenswrapper[4767]: I0317 15:43:16.563680 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rq699"] Mar 17 15:43:17 crc kubenswrapper[4767]: I0317 15:43:17.364051 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" path="/var/lib/kubelet/pods/3dc430ad-a22a-4fe8-a4dd-5c94622da680/volumes" Mar 17 15:43:17 crc kubenswrapper[4767]: I0317 15:43:17.365652 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" path="/var/lib/kubelet/pods/7d1d71c5-161c-4ba3-96da-ed07128ff6e1/volumes" Mar 17 15:43:17 crc kubenswrapper[4767]: I0317 15:43:17.366355 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" path="/var/lib/kubelet/pods/7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a/volumes" Mar 17 15:43:17 crc kubenswrapper[4767]: I0317 15:43:17.367679 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" path="/var/lib/kubelet/pods/e76337aa-429a-49b9-99d3-e261c32ee85b/volumes" Mar 17 15:43:17 crc kubenswrapper[4767]: I0317 15:43:17.368675 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" path="/var/lib/kubelet/pods/ebb17b29-f33c-4236-bb9d-034821b17ba3/volumes" Mar 17 15:43:17 crc kubenswrapper[4767]: I0317 15:43:17.370483 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" path="/var/lib/kubelet/pods/f46dfeb0-7bd5-4835-a16e-8a117a3db8a3/volumes" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.236474 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v5jdf"] Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237287 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237304 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237316 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237322 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237330 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237336 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237345 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237351 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237360 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237366 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237373 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237379 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237390 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237396 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237405 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237411 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237423 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237431 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237441 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237448 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237457 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237465 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237476 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237482 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237489 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237495 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237501 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237507 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237517 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237522 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237531 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acdeabaa-85fb-4d85-bad6-89fc1be0125a" containerName="marketplace-operator" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237536 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="acdeabaa-85fb-4d85-bad6-89fc1be0125a" containerName="marketplace-operator" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237543 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237549 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237557 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237563 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerName="extract-utilities" Mar 17 15:43:30 crc kubenswrapper[4767]: E0317 15:43:30.237570 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237576 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerName="extract-content" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237684 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc430ad-a22a-4fe8-a4dd-5c94622da680" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237699 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebb17b29-f33c-4236-bb9d-034821b17ba3" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237708 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="e76337aa-429a-49b9-99d3-e261c32ee85b" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237716 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d1d71c5-161c-4ba3-96da-ed07128ff6e1" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237723 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="acdeabaa-85fb-4d85-bad6-89fc1be0125a" containerName="marketplace-operator" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237729 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee8aaaa-9a6c-4ef6-9adb-b8f0ee50ee9a" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.237738 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f46dfeb0-7bd5-4835-a16e-8a117a3db8a3" containerName="registry-server" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.238663 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.242101 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.245578 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v5jdf"] Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.270975 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24b4b439-f309-4551-86d6-fc26f57fe754-utilities\") pod \"redhat-operators-v5jdf\" (UID: \"24b4b439-f309-4551-86d6-fc26f57fe754\") " pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.271023 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24b4b439-f309-4551-86d6-fc26f57fe754-catalog-content\") pod \"redhat-operators-v5jdf\" (UID: \"24b4b439-f309-4551-86d6-fc26f57fe754\") " pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.271094 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw9f6\" (UniqueName: \"kubernetes.io/projected/24b4b439-f309-4551-86d6-fc26f57fe754-kube-api-access-pw9f6\") pod \"redhat-operators-v5jdf\" (UID: \"24b4b439-f309-4551-86d6-fc26f57fe754\") " pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.372111 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw9f6\" (UniqueName: \"kubernetes.io/projected/24b4b439-f309-4551-86d6-fc26f57fe754-kube-api-access-pw9f6\") pod \"redhat-operators-v5jdf\" (UID: \"24b4b439-f309-4551-86d6-fc26f57fe754\") " pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.372185 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24b4b439-f309-4551-86d6-fc26f57fe754-utilities\") pod \"redhat-operators-v5jdf\" (UID: \"24b4b439-f309-4551-86d6-fc26f57fe754\") " pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.372213 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24b4b439-f309-4551-86d6-fc26f57fe754-catalog-content\") pod \"redhat-operators-v5jdf\" (UID: \"24b4b439-f309-4551-86d6-fc26f57fe754\") " pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.372718 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24b4b439-f309-4551-86d6-fc26f57fe754-utilities\") pod \"redhat-operators-v5jdf\" (UID: \"24b4b439-f309-4551-86d6-fc26f57fe754\") " pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.372841 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24b4b439-f309-4551-86d6-fc26f57fe754-catalog-content\") pod \"redhat-operators-v5jdf\" (UID: \"24b4b439-f309-4551-86d6-fc26f57fe754\") " pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.392579 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw9f6\" (UniqueName: \"kubernetes.io/projected/24b4b439-f309-4551-86d6-fc26f57fe754-kube-api-access-pw9f6\") pod \"redhat-operators-v5jdf\" (UID: \"24b4b439-f309-4551-86d6-fc26f57fe754\") " pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.555314 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:30 crc kubenswrapper[4767]: I0317 15:43:30.974057 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v5jdf"] Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.439093 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dsjwf"] Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.440578 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.442776 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.453302 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dsjwf"] Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.487561 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-catalog-content\") pod \"community-operators-dsjwf\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.487660 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk9kx\" (UniqueName: \"kubernetes.io/projected/7a3d6a4a-3683-4036-8171-b6488fc6b217-kube-api-access-vk9kx\") pod \"community-operators-dsjwf\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.487690 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-utilities\") pod \"community-operators-dsjwf\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.558395 4767 generic.go:334] "Generic (PLEG): container finished" podID="24b4b439-f309-4551-86d6-fc26f57fe754" containerID="3100d366d09665d72eb73e81869495de606ba17aaa1ab784bdc01e4ba6d41a66" exitCode=0 Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.558445 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5jdf" event={"ID":"24b4b439-f309-4551-86d6-fc26f57fe754","Type":"ContainerDied","Data":"3100d366d09665d72eb73e81869495de606ba17aaa1ab784bdc01e4ba6d41a66"} Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.558474 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5jdf" event={"ID":"24b4b439-f309-4551-86d6-fc26f57fe754","Type":"ContainerStarted","Data":"13367718999a964ea47ab5d7a648e435828b786ef4866b32fcb8b61c74724b89"} Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.588691 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk9kx\" (UniqueName: \"kubernetes.io/projected/7a3d6a4a-3683-4036-8171-b6488fc6b217-kube-api-access-vk9kx\") pod \"community-operators-dsjwf\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.588729 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-utilities\") pod \"community-operators-dsjwf\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.588786 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-catalog-content\") pod \"community-operators-dsjwf\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.589209 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-catalog-content\") pod \"community-operators-dsjwf\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.589283 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-utilities\") pod \"community-operators-dsjwf\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.606623 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk9kx\" (UniqueName: \"kubernetes.io/projected/7a3d6a4a-3683-4036-8171-b6488fc6b217-kube-api-access-vk9kx\") pod \"community-operators-dsjwf\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:31 crc kubenswrapper[4767]: I0317 15:43:31.798053 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:32 crc kubenswrapper[4767]: I0317 15:43:32.022122 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dsjwf"] Mar 17 15:43:32 crc kubenswrapper[4767]: W0317 15:43:32.028584 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a3d6a4a_3683_4036_8171_b6488fc6b217.slice/crio-7f00cbc6d4c00b82838315af6b66de59c80fc66a75529c8493a06c592e9f9aed WatchSource:0}: Error finding container 7f00cbc6d4c00b82838315af6b66de59c80fc66a75529c8493a06c592e9f9aed: Status 404 returned error can't find the container with id 7f00cbc6d4c00b82838315af6b66de59c80fc66a75529c8493a06c592e9f9aed Mar 17 15:43:32 crc kubenswrapper[4767]: I0317 15:43:32.566535 4767 generic.go:334] "Generic (PLEG): container finished" podID="7a3d6a4a-3683-4036-8171-b6488fc6b217" containerID="7ec299dce77cb11f7c42ff2808ec238639cce0d064504cca4f2b55430367a5ea" exitCode=0 Mar 17 15:43:32 crc kubenswrapper[4767]: I0317 15:43:32.566648 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dsjwf" event={"ID":"7a3d6a4a-3683-4036-8171-b6488fc6b217","Type":"ContainerDied","Data":"7ec299dce77cb11f7c42ff2808ec238639cce0d064504cca4f2b55430367a5ea"} Mar 17 15:43:32 crc kubenswrapper[4767]: I0317 15:43:32.566695 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dsjwf" event={"ID":"7a3d6a4a-3683-4036-8171-b6488fc6b217","Type":"ContainerStarted","Data":"7f00cbc6d4c00b82838315af6b66de59c80fc66a75529c8493a06c592e9f9aed"} Mar 17 15:43:32 crc kubenswrapper[4767]: I0317 15:43:32.570558 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5jdf" event={"ID":"24b4b439-f309-4551-86d6-fc26f57fe754","Type":"ContainerStarted","Data":"895a886aefd6012848c32d41d507d6a63be2aa781d95d0cbc30cf6a8611e3691"} Mar 17 15:43:32 crc kubenswrapper[4767]: I0317 15:43:32.837312 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wnf5c"] Mar 17 15:43:32 crc kubenswrapper[4767]: I0317 15:43:32.839557 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:32 crc kubenswrapper[4767]: I0317 15:43:32.842499 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 17 15:43:32 crc kubenswrapper[4767]: I0317 15:43:32.854201 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wnf5c"] Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.005728 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-utilities\") pod \"certified-operators-wnf5c\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.006025 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-catalog-content\") pod \"certified-operators-wnf5c\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.006473 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnpt5\" (UniqueName: \"kubernetes.io/projected/25c1b139-cd56-4128-ac94-9c0b537f0c0d-kube-api-access-hnpt5\") pod \"certified-operators-wnf5c\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.108494 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-utilities\") pod \"certified-operators-wnf5c\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.108570 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-catalog-content\") pod \"certified-operators-wnf5c\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.108647 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnpt5\" (UniqueName: \"kubernetes.io/projected/25c1b139-cd56-4128-ac94-9c0b537f0c0d-kube-api-access-hnpt5\") pod \"certified-operators-wnf5c\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.109080 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-utilities\") pod \"certified-operators-wnf5c\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.109592 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-catalog-content\") pod \"certified-operators-wnf5c\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.129087 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnpt5\" (UniqueName: \"kubernetes.io/projected/25c1b139-cd56-4128-ac94-9c0b537f0c0d-kube-api-access-hnpt5\") pod \"certified-operators-wnf5c\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.175736 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.849973 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sl4m8"] Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.851310 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.857583 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.863122 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sl4m8"] Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.900662 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a3dbff-bd98-4264-974a-5c098c6f1361-utilities\") pod \"redhat-marketplace-sl4m8\" (UID: \"56a3dbff-bd98-4264-974a-5c098c6f1361\") " pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.900716 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a3dbff-bd98-4264-974a-5c098c6f1361-catalog-content\") pod \"redhat-marketplace-sl4m8\" (UID: \"56a3dbff-bd98-4264-974a-5c098c6f1361\") " pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:33 crc kubenswrapper[4767]: I0317 15:43:33.900823 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqhxq\" (UniqueName: \"kubernetes.io/projected/56a3dbff-bd98-4264-974a-5c098c6f1361-kube-api-access-bqhxq\") pod \"redhat-marketplace-sl4m8\" (UID: \"56a3dbff-bd98-4264-974a-5c098c6f1361\") " pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.001761 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a3dbff-bd98-4264-974a-5c098c6f1361-utilities\") pod \"redhat-marketplace-sl4m8\" (UID: \"56a3dbff-bd98-4264-974a-5c098c6f1361\") " pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.001816 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a3dbff-bd98-4264-974a-5c098c6f1361-catalog-content\") pod \"redhat-marketplace-sl4m8\" (UID: \"56a3dbff-bd98-4264-974a-5c098c6f1361\") " pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.001864 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqhxq\" (UniqueName: \"kubernetes.io/projected/56a3dbff-bd98-4264-974a-5c098c6f1361-kube-api-access-bqhxq\") pod \"redhat-marketplace-sl4m8\" (UID: \"56a3dbff-bd98-4264-974a-5c098c6f1361\") " pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.002243 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a3dbff-bd98-4264-974a-5c098c6f1361-utilities\") pod \"redhat-marketplace-sl4m8\" (UID: \"56a3dbff-bd98-4264-974a-5c098c6f1361\") " pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.002607 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a3dbff-bd98-4264-974a-5c098c6f1361-catalog-content\") pod \"redhat-marketplace-sl4m8\" (UID: \"56a3dbff-bd98-4264-974a-5c098c6f1361\") " pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.023644 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqhxq\" (UniqueName: \"kubernetes.io/projected/56a3dbff-bd98-4264-974a-5c098c6f1361-kube-api-access-bqhxq\") pod \"redhat-marketplace-sl4m8\" (UID: \"56a3dbff-bd98-4264-974a-5c098c6f1361\") " pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.178047 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.241627 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wnf5c"] Mar 17 15:43:34 crc kubenswrapper[4767]: W0317 15:43:34.244912 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25c1b139_cd56_4128_ac94_9c0b537f0c0d.slice/crio-36e4d9531b81bcc28adf6e2021ac0ebb1eb4dcf767078cc24e2a922c0c3afabf WatchSource:0}: Error finding container 36e4d9531b81bcc28adf6e2021ac0ebb1eb4dcf767078cc24e2a922c0c3afabf: Status 404 returned error can't find the container with id 36e4d9531b81bcc28adf6e2021ac0ebb1eb4dcf767078cc24e2a922c0c3afabf Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.939447 4767 generic.go:334] "Generic (PLEG): container finished" podID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" containerID="fa047f500d41d4af3255fb1beb5ea59580f132d6384a445aad7338207becee0c" exitCode=0 Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.939526 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnf5c" event={"ID":"25c1b139-cd56-4128-ac94-9c0b537f0c0d","Type":"ContainerDied","Data":"fa047f500d41d4af3255fb1beb5ea59580f132d6384a445aad7338207becee0c"} Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.939550 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnf5c" event={"ID":"25c1b139-cd56-4128-ac94-9c0b537f0c0d","Type":"ContainerStarted","Data":"36e4d9531b81bcc28adf6e2021ac0ebb1eb4dcf767078cc24e2a922c0c3afabf"} Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.941388 4767 generic.go:334] "Generic (PLEG): container finished" podID="24b4b439-f309-4551-86d6-fc26f57fe754" containerID="895a886aefd6012848c32d41d507d6a63be2aa781d95d0cbc30cf6a8611e3691" exitCode=0 Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.941469 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5jdf" event={"ID":"24b4b439-f309-4551-86d6-fc26f57fe754","Type":"ContainerDied","Data":"895a886aefd6012848c32d41d507d6a63be2aa781d95d0cbc30cf6a8611e3691"} Mar 17 15:43:34 crc kubenswrapper[4767]: I0317 15:43:34.954069 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sl4m8"] Mar 17 15:43:35 crc kubenswrapper[4767]: I0317 15:43:35.950565 4767 generic.go:334] "Generic (PLEG): container finished" podID="7a3d6a4a-3683-4036-8171-b6488fc6b217" containerID="f92eb9a109ebf9102ec71a27648137d8e8b5fa2b290dc68109614b9dc94dfbff" exitCode=0 Mar 17 15:43:35 crc kubenswrapper[4767]: I0317 15:43:35.950627 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dsjwf" event={"ID":"7a3d6a4a-3683-4036-8171-b6488fc6b217","Type":"ContainerDied","Data":"f92eb9a109ebf9102ec71a27648137d8e8b5fa2b290dc68109614b9dc94dfbff"} Mar 17 15:43:35 crc kubenswrapper[4767]: I0317 15:43:35.955865 4767 generic.go:334] "Generic (PLEG): container finished" podID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerID="0b359572392233c0ecf84895c02c6b344c3b2a4aa77710fdaca19ed26646f046" exitCode=0 Mar 17 15:43:35 crc kubenswrapper[4767]: I0317 15:43:35.955910 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl4m8" event={"ID":"56a3dbff-bd98-4264-974a-5c098c6f1361","Type":"ContainerDied","Data":"0b359572392233c0ecf84895c02c6b344c3b2a4aa77710fdaca19ed26646f046"} Mar 17 15:43:35 crc kubenswrapper[4767]: I0317 15:43:35.955937 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl4m8" event={"ID":"56a3dbff-bd98-4264-974a-5c098c6f1361","Type":"ContainerStarted","Data":"7da6d69f857b73cdf20b75e33e459afa5a0e97635e942ebdb09f6adfdd650796"} Mar 17 15:43:36 crc kubenswrapper[4767]: I0317 15:43:36.963226 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dsjwf" event={"ID":"7a3d6a4a-3683-4036-8171-b6488fc6b217","Type":"ContainerStarted","Data":"31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd"} Mar 17 15:43:36 crc kubenswrapper[4767]: I0317 15:43:36.965554 4767 generic.go:334] "Generic (PLEG): container finished" podID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" containerID="7531c28fdd615c44e17b1414549a58e4c9498789ecb3364760c2db5569c5f29d" exitCode=0 Mar 17 15:43:36 crc kubenswrapper[4767]: I0317 15:43:36.965650 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnf5c" event={"ID":"25c1b139-cd56-4128-ac94-9c0b537f0c0d","Type":"ContainerDied","Data":"7531c28fdd615c44e17b1414549a58e4c9498789ecb3364760c2db5569c5f29d"} Mar 17 15:43:36 crc kubenswrapper[4767]: I0317 15:43:36.968353 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5jdf" event={"ID":"24b4b439-f309-4551-86d6-fc26f57fe754","Type":"ContainerStarted","Data":"44e01702cdd6851f3e6eac45406e1af0302eab03207b5ef699665002ceb62b55"} Mar 17 15:43:36 crc kubenswrapper[4767]: I0317 15:43:36.972003 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl4m8" event={"ID":"56a3dbff-bd98-4264-974a-5c098c6f1361","Type":"ContainerStarted","Data":"479643bd84cfc7f745cc0b7a05f8c740f273adfa3436dbfb5324ea572c623f9f"} Mar 17 15:43:36 crc kubenswrapper[4767]: I0317 15:43:36.989058 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dsjwf" podStartSLOduration=2.14147321 podStartE2EDuration="5.989038477s" podCreationTimestamp="2026-03-17 15:43:31 +0000 UTC" firstStartedPulling="2026-03-17 15:43:32.568829089 +0000 UTC m=+403.982145136" lastFinishedPulling="2026-03-17 15:43:36.416394356 +0000 UTC m=+407.829710403" observedRunningTime="2026-03-17 15:43:36.98619111 +0000 UTC m=+408.399507167" watchObservedRunningTime="2026-03-17 15:43:36.989038477 +0000 UTC m=+408.402354524" Mar 17 15:43:38 crc kubenswrapper[4767]: I0317 15:43:38.146384 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnf5c" event={"ID":"25c1b139-cd56-4128-ac94-9c0b537f0c0d","Type":"ContainerStarted","Data":"f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512"} Mar 17 15:43:38 crc kubenswrapper[4767]: I0317 15:43:38.156285 4767 generic.go:334] "Generic (PLEG): container finished" podID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerID="479643bd84cfc7f745cc0b7a05f8c740f273adfa3436dbfb5324ea572c623f9f" exitCode=0 Mar 17 15:43:38 crc kubenswrapper[4767]: I0317 15:43:38.157367 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl4m8" event={"ID":"56a3dbff-bd98-4264-974a-5c098c6f1361","Type":"ContainerDied","Data":"479643bd84cfc7f745cc0b7a05f8c740f273adfa3436dbfb5324ea572c623f9f"} Mar 17 15:43:38 crc kubenswrapper[4767]: I0317 15:43:38.174934 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v5jdf" podStartSLOduration=4.09100169 podStartE2EDuration="8.174912785s" podCreationTimestamp="2026-03-17 15:43:30 +0000 UTC" firstStartedPulling="2026-03-17 15:43:31.559865468 +0000 UTC m=+402.973181515" lastFinishedPulling="2026-03-17 15:43:35.643776543 +0000 UTC m=+407.057092610" observedRunningTime="2026-03-17 15:43:37.216784893 +0000 UTC m=+408.630100950" watchObservedRunningTime="2026-03-17 15:43:38.174912785 +0000 UTC m=+409.588228832" Mar 17 15:43:38 crc kubenswrapper[4767]: I0317 15:43:38.199812 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wnf5c" podStartSLOduration=3.638780722 podStartE2EDuration="6.199793733s" podCreationTimestamp="2026-03-17 15:43:32 +0000 UTC" firstStartedPulling="2026-03-17 15:43:34.941624448 +0000 UTC m=+406.354940495" lastFinishedPulling="2026-03-17 15:43:37.502637459 +0000 UTC m=+408.915953506" observedRunningTime="2026-03-17 15:43:38.180065072 +0000 UTC m=+409.593381149" watchObservedRunningTime="2026-03-17 15:43:38.199793733 +0000 UTC m=+409.613109770" Mar 17 15:43:39 crc kubenswrapper[4767]: I0317 15:43:39.168926 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl4m8" event={"ID":"56a3dbff-bd98-4264-974a-5c098c6f1361","Type":"ContainerStarted","Data":"80cdce19b94103bc1982d3760e2748cdc19f5e1e69691384f05bd4ed1d8cf95e"} Mar 17 15:43:40 crc kubenswrapper[4767]: I0317 15:43:40.559622 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:40 crc kubenswrapper[4767]: I0317 15:43:40.563186 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:41 crc kubenswrapper[4767]: I0317 15:43:41.608331 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output=< Mar 17 15:43:41 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 15:43:41 crc kubenswrapper[4767]: > Mar 17 15:43:41 crc kubenswrapper[4767]: I0317 15:43:41.798564 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:41 crc kubenswrapper[4767]: I0317 15:43:41.800464 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:41 crc kubenswrapper[4767]: I0317 15:43:41.842099 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:41 crc kubenswrapper[4767]: I0317 15:43:41.864378 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sl4m8" podStartSLOduration=6.232995851 podStartE2EDuration="8.864352955s" podCreationTimestamp="2026-03-17 15:43:33 +0000 UTC" firstStartedPulling="2026-03-17 15:43:35.957238441 +0000 UTC m=+407.370554488" lastFinishedPulling="2026-03-17 15:43:38.588595545 +0000 UTC m=+410.001911592" observedRunningTime="2026-03-17 15:43:39.195683455 +0000 UTC m=+410.608999502" watchObservedRunningTime="2026-03-17 15:43:41.864352955 +0000 UTC m=+413.277669022" Mar 17 15:43:42 crc kubenswrapper[4767]: I0317 15:43:42.307111 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dsjwf" Mar 17 15:43:43 crc kubenswrapper[4767]: I0317 15:43:43.175957 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:43 crc kubenswrapper[4767]: I0317 15:43:43.176350 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:43 crc kubenswrapper[4767]: I0317 15:43:43.217976 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:43 crc kubenswrapper[4767]: I0317 15:43:43.302784 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 15:43:44 crc kubenswrapper[4767]: I0317 15:43:44.179831 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:44 crc kubenswrapper[4767]: I0317 15:43:44.180406 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:44 crc kubenswrapper[4767]: I0317 15:43:44.223199 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:44 crc kubenswrapper[4767]: I0317 15:43:44.299571 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 15:43:50 crc kubenswrapper[4767]: I0317 15:43:50.822475 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:43:50 crc kubenswrapper[4767]: I0317 15:43:50.872919 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 15:44:00 crc kubenswrapper[4767]: I0317 15:44:00.202425 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562704-kc8jt"] Mar 17 15:44:00 crc kubenswrapper[4767]: I0317 15:44:00.205418 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562704-kc8jt" Mar 17 15:44:00 crc kubenswrapper[4767]: I0317 15:44:00.208368 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 15:44:00 crc kubenswrapper[4767]: I0317 15:44:00.208438 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 15:44:00 crc kubenswrapper[4767]: I0317 15:44:00.211250 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 15:44:00 crc kubenswrapper[4767]: I0317 15:44:00.228379 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562704-kc8jt"] Mar 17 15:44:00 crc kubenswrapper[4767]: I0317 15:44:00.344535 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ngjj\" (UniqueName: \"kubernetes.io/projected/6747c3a7-dc68-4728-bbe4-f642dae2d4b0-kube-api-access-7ngjj\") pod \"auto-csr-approver-29562704-kc8jt\" (UID: \"6747c3a7-dc68-4728-bbe4-f642dae2d4b0\") " pod="openshift-infra/auto-csr-approver-29562704-kc8jt" Mar 17 15:44:00 crc kubenswrapper[4767]: I0317 15:44:00.445546 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ngjj\" (UniqueName: \"kubernetes.io/projected/6747c3a7-dc68-4728-bbe4-f642dae2d4b0-kube-api-access-7ngjj\") pod \"auto-csr-approver-29562704-kc8jt\" (UID: \"6747c3a7-dc68-4728-bbe4-f642dae2d4b0\") " pod="openshift-infra/auto-csr-approver-29562704-kc8jt" Mar 17 15:44:00 crc kubenswrapper[4767]: I0317 15:44:00.468231 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ngjj\" (UniqueName: \"kubernetes.io/projected/6747c3a7-dc68-4728-bbe4-f642dae2d4b0-kube-api-access-7ngjj\") pod \"auto-csr-approver-29562704-kc8jt\" (UID: \"6747c3a7-dc68-4728-bbe4-f642dae2d4b0\") " pod="openshift-infra/auto-csr-approver-29562704-kc8jt" Mar 17 15:44:00 crc kubenswrapper[4767]: I0317 15:44:00.531533 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562704-kc8jt" Mar 17 15:44:01 crc kubenswrapper[4767]: I0317 15:44:01.038873 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562704-kc8jt"] Mar 17 15:44:01 crc kubenswrapper[4767]: I0317 15:44:01.197998 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562704-kc8jt" event={"ID":"6747c3a7-dc68-4728-bbe4-f642dae2d4b0","Type":"ContainerStarted","Data":"ccbdceb11d3618ed1aea3c2fc4124fc8c331f8bbdaf25b48f0d194d7d6ea35d6"} Mar 17 15:44:03 crc kubenswrapper[4767]: I0317 15:44:03.210133 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562704-kc8jt" event={"ID":"6747c3a7-dc68-4728-bbe4-f642dae2d4b0","Type":"ContainerStarted","Data":"22f7c90dd5848ad907327e3b3134bde167e737517b6f81ee131d64d6080bea08"} Mar 17 15:44:03 crc kubenswrapper[4767]: I0317 15:44:03.235855 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562704-kc8jt" podStartSLOduration=1.732780924 podStartE2EDuration="3.235824282s" podCreationTimestamp="2026-03-17 15:44:00 +0000 UTC" firstStartedPulling="2026-03-17 15:44:01.047310686 +0000 UTC m=+432.460626733" lastFinishedPulling="2026-03-17 15:44:02.550354054 +0000 UTC m=+433.963670091" observedRunningTime="2026-03-17 15:44:03.232191911 +0000 UTC m=+434.645507978" watchObservedRunningTime="2026-03-17 15:44:03.235824282 +0000 UTC m=+434.649140349" Mar 17 15:44:04 crc kubenswrapper[4767]: I0317 15:44:04.167199 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:44:04 crc kubenswrapper[4767]: I0317 15:44:04.167558 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:44:04 crc kubenswrapper[4767]: I0317 15:44:04.221531 4767 generic.go:334] "Generic (PLEG): container finished" podID="6747c3a7-dc68-4728-bbe4-f642dae2d4b0" containerID="22f7c90dd5848ad907327e3b3134bde167e737517b6f81ee131d64d6080bea08" exitCode=0 Mar 17 15:44:04 crc kubenswrapper[4767]: I0317 15:44:04.221598 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562704-kc8jt" event={"ID":"6747c3a7-dc68-4728-bbe4-f642dae2d4b0","Type":"ContainerDied","Data":"22f7c90dd5848ad907327e3b3134bde167e737517b6f81ee131d64d6080bea08"} Mar 17 15:44:05 crc kubenswrapper[4767]: I0317 15:44:05.584093 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562704-kc8jt" Mar 17 15:44:05 crc kubenswrapper[4767]: I0317 15:44:05.628009 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ngjj\" (UniqueName: \"kubernetes.io/projected/6747c3a7-dc68-4728-bbe4-f642dae2d4b0-kube-api-access-7ngjj\") pod \"6747c3a7-dc68-4728-bbe4-f642dae2d4b0\" (UID: \"6747c3a7-dc68-4728-bbe4-f642dae2d4b0\") " Mar 17 15:44:05 crc kubenswrapper[4767]: I0317 15:44:05.635643 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6747c3a7-dc68-4728-bbe4-f642dae2d4b0-kube-api-access-7ngjj" (OuterVolumeSpecName: "kube-api-access-7ngjj") pod "6747c3a7-dc68-4728-bbe4-f642dae2d4b0" (UID: "6747c3a7-dc68-4728-bbe4-f642dae2d4b0"). InnerVolumeSpecName "kube-api-access-7ngjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:44:05 crc kubenswrapper[4767]: I0317 15:44:05.730025 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ngjj\" (UniqueName: \"kubernetes.io/projected/6747c3a7-dc68-4728-bbe4-f642dae2d4b0-kube-api-access-7ngjj\") on node \"crc\" DevicePath \"\"" Mar 17 15:44:06 crc kubenswrapper[4767]: I0317 15:44:06.235526 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562704-kc8jt" event={"ID":"6747c3a7-dc68-4728-bbe4-f642dae2d4b0","Type":"ContainerDied","Data":"ccbdceb11d3618ed1aea3c2fc4124fc8c331f8bbdaf25b48f0d194d7d6ea35d6"} Mar 17 15:44:06 crc kubenswrapper[4767]: I0317 15:44:06.235593 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccbdceb11d3618ed1aea3c2fc4124fc8c331f8bbdaf25b48f0d194d7d6ea35d6" Mar 17 15:44:06 crc kubenswrapper[4767]: I0317 15:44:06.235627 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562704-kc8jt" Mar 17 15:44:12 crc kubenswrapper[4767]: I0317 15:44:12.996874 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd"] Mar 17 15:44:12 crc kubenswrapper[4767]: E0317 15:44:12.998289 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6747c3a7-dc68-4728-bbe4-f642dae2d4b0" containerName="oc" Mar 17 15:44:12 crc kubenswrapper[4767]: I0317 15:44:12.998316 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6747c3a7-dc68-4728-bbe4-f642dae2d4b0" containerName="oc" Mar 17 15:44:12 crc kubenswrapper[4767]: I0317 15:44:12.998522 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="6747c3a7-dc68-4728-bbe4-f642dae2d4b0" containerName="oc" Mar 17 15:44:12 crc kubenswrapper[4767]: I0317 15:44:12.999628 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.003292 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.003558 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.005811 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.006562 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.007539 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.018784 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd"] Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.074119 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9pbn\" (UniqueName: \"kubernetes.io/projected/1a25d556-1468-4fb9-8697-b6d1107dc3ed-kube-api-access-x9pbn\") pod \"cluster-monitoring-operator-6d5b84845-tsrhd\" (UID: \"1a25d556-1468-4fb9-8697-b6d1107dc3ed\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.074272 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a25d556-1468-4fb9-8697-b6d1107dc3ed-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-tsrhd\" (UID: \"1a25d556-1468-4fb9-8697-b6d1107dc3ed\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.074347 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/1a25d556-1468-4fb9-8697-b6d1107dc3ed-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-tsrhd\" (UID: \"1a25d556-1468-4fb9-8697-b6d1107dc3ed\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.175564 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/1a25d556-1468-4fb9-8697-b6d1107dc3ed-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-tsrhd\" (UID: \"1a25d556-1468-4fb9-8697-b6d1107dc3ed\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.175658 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9pbn\" (UniqueName: \"kubernetes.io/projected/1a25d556-1468-4fb9-8697-b6d1107dc3ed-kube-api-access-x9pbn\") pod \"cluster-monitoring-operator-6d5b84845-tsrhd\" (UID: \"1a25d556-1468-4fb9-8697-b6d1107dc3ed\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.175711 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a25d556-1468-4fb9-8697-b6d1107dc3ed-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-tsrhd\" (UID: \"1a25d556-1468-4fb9-8697-b6d1107dc3ed\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.176836 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/1a25d556-1468-4fb9-8697-b6d1107dc3ed-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-tsrhd\" (UID: \"1a25d556-1468-4fb9-8697-b6d1107dc3ed\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.189686 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a25d556-1468-4fb9-8697-b6d1107dc3ed-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-tsrhd\" (UID: \"1a25d556-1468-4fb9-8697-b6d1107dc3ed\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.197710 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9pbn\" (UniqueName: \"kubernetes.io/projected/1a25d556-1468-4fb9-8697-b6d1107dc3ed-kube-api-access-x9pbn\") pod \"cluster-monitoring-operator-6d5b84845-tsrhd\" (UID: \"1a25d556-1468-4fb9-8697-b6d1107dc3ed\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.327116 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" Mar 17 15:44:13 crc kubenswrapper[4767]: I0317 15:44:13.872112 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd"] Mar 17 15:44:14 crc kubenswrapper[4767]: I0317 15:44:14.299438 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" event={"ID":"1a25d556-1468-4fb9-8697-b6d1107dc3ed","Type":"ContainerStarted","Data":"39eb53571b4658c6f577949aafb0639b352e822983a16ca7f91cef7b0fb77fbe"} Mar 17 15:44:17 crc kubenswrapper[4767]: I0317 15:44:17.318937 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" event={"ID":"1a25d556-1468-4fb9-8697-b6d1107dc3ed","Type":"ContainerStarted","Data":"f0004fba665ed51e9f0a2b7f7914ef13de29ebc039d22793fad31c19565ba6da"} Mar 17 15:44:17 crc kubenswrapper[4767]: I0317 15:44:17.340252 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-tsrhd" podStartSLOduration=2.29810578 podStartE2EDuration="5.340226284s" podCreationTimestamp="2026-03-17 15:44:12 +0000 UTC" firstStartedPulling="2026-03-17 15:44:13.888266857 +0000 UTC m=+445.301582904" lastFinishedPulling="2026-03-17 15:44:16.930387361 +0000 UTC m=+448.343703408" observedRunningTime="2026-03-17 15:44:17.334092907 +0000 UTC m=+448.747408984" watchObservedRunningTime="2026-03-17 15:44:17.340226284 +0000 UTC m=+448.753542351" Mar 17 15:44:19 crc kubenswrapper[4767]: I0317 15:44:19.074263 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b"] Mar 17 15:44:19 crc kubenswrapper[4767]: I0317 15:44:19.076095 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 15:44:19 crc kubenswrapper[4767]: I0317 15:44:19.082316 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-dockercfg-szj7z" Mar 17 15:44:19 crc kubenswrapper[4767]: I0317 15:44:19.094642 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Mar 17 15:44:19 crc kubenswrapper[4767]: I0317 15:44:19.125679 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b"] Mar 17 15:44:19 crc kubenswrapper[4767]: I0317 15:44:19.788193 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-82k6b\" (UID: \"15e74b29-0b89-4e6a-84aa-94d9f5faa0e5\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 15:44:19 crc kubenswrapper[4767]: I0317 15:44:19.890055 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-82k6b\" (UID: \"15e74b29-0b89-4e6a-84aa-94d9f5faa0e5\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 15:44:19 crc kubenswrapper[4767]: E0317 15:44:19.890485 4767 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Mar 17 15:44:19 crc kubenswrapper[4767]: E0317 15:44:19.890605 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates podName:15e74b29-0b89-4e6a-84aa-94d9f5faa0e5 nodeName:}" failed. No retries permitted until 2026-03-17 15:44:20.3905863 +0000 UTC m=+451.803902347 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-82k6b" (UID: "15e74b29-0b89-4e6a-84aa-94d9f5faa0e5") : secret "prometheus-operator-admission-webhook-tls" not found Mar 17 15:44:19 crc kubenswrapper[4767]: I0317 15:44:19.898224 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j7tgx"] Mar 17 15:44:19 crc kubenswrapper[4767]: I0317 15:44:19.899405 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:19 crc kubenswrapper[4767]: I0317 15:44:19.937908 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j7tgx"] Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.093034 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.093095 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.093122 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-registry-certificates\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.093144 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plgmr\" (UniqueName: \"kubernetes.io/projected/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-kube-api-access-plgmr\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.093191 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-bound-sa-token\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.093229 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.093274 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-trusted-ca\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.093308 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-registry-tls\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.117399 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.194607 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.194689 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-registry-certificates\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.194716 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plgmr\" (UniqueName: \"kubernetes.io/projected/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-kube-api-access-plgmr\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.194751 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-bound-sa-token\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.194782 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.194824 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-trusted-ca\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.194848 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-registry-tls\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.196485 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.196557 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-trusted-ca\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.196798 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-registry-certificates\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.204009 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.204064 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-registry-tls\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.217366 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-bound-sa-token\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.218617 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plgmr\" (UniqueName: \"kubernetes.io/projected/54f1fbaa-b95b-40ef-a7fa-89bb6efbc128-kube-api-access-plgmr\") pod \"image-registry-66df7c8f76-j7tgx\" (UID: \"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128\") " pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.229944 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.398420 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-82k6b\" (UID: \"15e74b29-0b89-4e6a-84aa-94d9f5faa0e5\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 15:44:20 crc kubenswrapper[4767]: E0317 15:44:20.398742 4767 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Mar 17 15:44:20 crc kubenswrapper[4767]: E0317 15:44:20.399010 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates podName:15e74b29-0b89-4e6a-84aa-94d9f5faa0e5 nodeName:}" failed. No retries permitted until 2026-03-17 15:44:21.398985985 +0000 UTC m=+452.812302032 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-82k6b" (UID: "15e74b29-0b89-4e6a-84aa-94d9f5faa0e5") : secret "prometheus-operator-admission-webhook-tls" not found Mar 17 15:44:20 crc kubenswrapper[4767]: I0317 15:44:20.702898 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j7tgx"] Mar 17 15:44:20 crc kubenswrapper[4767]: W0317 15:44:20.718647 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54f1fbaa_b95b_40ef_a7fa_89bb6efbc128.slice/crio-3f65d0dd37c594c0ff024a3f3222de6054a9d7b150a454eea76ca825237fde1d WatchSource:0}: Error finding container 3f65d0dd37c594c0ff024a3f3222de6054a9d7b150a454eea76ca825237fde1d: Status 404 returned error can't find the container with id 3f65d0dd37c594c0ff024a3f3222de6054a9d7b150a454eea76ca825237fde1d Mar 17 15:44:21 crc kubenswrapper[4767]: I0317 15:44:21.033706 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" event={"ID":"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128","Type":"ContainerStarted","Data":"3f65d0dd37c594c0ff024a3f3222de6054a9d7b150a454eea76ca825237fde1d"} Mar 17 15:44:21 crc kubenswrapper[4767]: I0317 15:44:21.534826 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-82k6b\" (UID: \"15e74b29-0b89-4e6a-84aa-94d9f5faa0e5\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 15:44:21 crc kubenswrapper[4767]: E0317 15:44:21.535017 4767 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Mar 17 15:44:21 crc kubenswrapper[4767]: E0317 15:44:21.535091 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates podName:15e74b29-0b89-4e6a-84aa-94d9f5faa0e5 nodeName:}" failed. No retries permitted until 2026-03-17 15:44:23.535068327 +0000 UTC m=+454.948384364 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-82k6b" (UID: "15e74b29-0b89-4e6a-84aa-94d9f5faa0e5") : secret "prometheus-operator-admission-webhook-tls" not found Mar 17 15:44:22 crc kubenswrapper[4767]: I0317 15:44:22.043247 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" event={"ID":"54f1fbaa-b95b-40ef-a7fa-89bb6efbc128","Type":"ContainerStarted","Data":"111478be7ba99052ef976bfd91a662bb994c11f8ddee683c38e18fe248962921"} Mar 17 15:44:22 crc kubenswrapper[4767]: I0317 15:44:22.043531 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:22 crc kubenswrapper[4767]: I0317 15:44:22.079873 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" podStartSLOduration=3.07984569 podStartE2EDuration="3.07984569s" podCreationTimestamp="2026-03-17 15:44:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:44:22.075145557 +0000 UTC m=+453.488461614" watchObservedRunningTime="2026-03-17 15:44:22.07984569 +0000 UTC m=+453.493161737" Mar 17 15:44:23 crc kubenswrapper[4767]: I0317 15:44:23.547972 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-82k6b\" (UID: \"15e74b29-0b89-4e6a-84aa-94d9f5faa0e5\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 15:44:23 crc kubenswrapper[4767]: E0317 15:44:23.548295 4767 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Mar 17 15:44:23 crc kubenswrapper[4767]: E0317 15:44:23.548395 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates podName:15e74b29-0b89-4e6a-84aa-94d9f5faa0e5 nodeName:}" failed. No retries permitted until 2026-03-17 15:44:27.548366198 +0000 UTC m=+458.961682245 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-82k6b" (UID: "15e74b29-0b89-4e6a-84aa-94d9f5faa0e5") : secret "prometheus-operator-admission-webhook-tls" not found Mar 17 15:44:27 crc kubenswrapper[4767]: I0317 15:44:27.562928 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-82k6b\" (UID: \"15e74b29-0b89-4e6a-84aa-94d9f5faa0e5\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 15:44:27 crc kubenswrapper[4767]: I0317 15:44:27.576279 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/15e74b29-0b89-4e6a-84aa-94d9f5faa0e5-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-82k6b\" (UID: \"15e74b29-0b89-4e6a-84aa-94d9f5faa0e5\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 15:44:27 crc kubenswrapper[4767]: I0317 15:44:27.830705 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 15:44:28 crc kubenswrapper[4767]: I0317 15:44:28.075720 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b"] Mar 17 15:44:28 crc kubenswrapper[4767]: I0317 15:44:28.689010 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" event={"ID":"15e74b29-0b89-4e6a-84aa-94d9f5faa0e5","Type":"ContainerStarted","Data":"029680008e24bc5c3f4534e56490361d5482728160fb84cd84f76cd97990c4c3"} Mar 17 15:44:30 crc kubenswrapper[4767]: I0317 15:44:30.714487 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" event={"ID":"15e74b29-0b89-4e6a-84aa-94d9f5faa0e5","Type":"ContainerStarted","Data":"21bd32970bf467e403774e6b15ad900b53bef84f68891cc87f2251b74697e862"} Mar 17 15:44:30 crc kubenswrapper[4767]: I0317 15:44:30.715993 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 15:44:30 crc kubenswrapper[4767]: I0317 15:44:30.722275 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 15:44:30 crc kubenswrapper[4767]: I0317 15:44:30.737149 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podStartSLOduration=12.11792355 podStartE2EDuration="13.737116047s" podCreationTimestamp="2026-03-17 15:44:17 +0000 UTC" firstStartedPulling="2026-03-17 15:44:28.095267842 +0000 UTC m=+459.508583889" lastFinishedPulling="2026-03-17 15:44:29.714460339 +0000 UTC m=+461.127776386" observedRunningTime="2026-03-17 15:44:30.73326534 +0000 UTC m=+462.146581397" watchObservedRunningTime="2026-03-17 15:44:30.737116047 +0000 UTC m=+462.150432094" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.693132 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-zmvx6"] Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.694571 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.697049 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-6hqnz" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.698239 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.698711 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.699255 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.713084 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-zmvx6"] Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.853673 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-metrics-client-ca\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.854995 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.855199 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7xx9\" (UniqueName: \"kubernetes.io/projected/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-kube-api-access-v7xx9\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.855396 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.957732 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.957832 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7xx9\" (UniqueName: \"kubernetes.io/projected/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-kube-api-access-v7xx9\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.957904 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.958017 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-metrics-client-ca\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.960368 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-metrics-client-ca\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.967037 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.967577 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:31 crc kubenswrapper[4767]: I0317 15:44:31.981721 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7xx9\" (UniqueName: \"kubernetes.io/projected/b1521838-7bb4-44a8-9ceb-758a3fbb6ae1-kube-api-access-v7xx9\") pod \"prometheus-operator-db54df47d-zmvx6\" (UID: \"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:32 crc kubenswrapper[4767]: I0317 15:44:32.016761 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" Mar 17 15:44:32 crc kubenswrapper[4767]: I0317 15:44:32.345069 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-zmvx6"] Mar 17 15:44:32 crc kubenswrapper[4767]: I0317 15:44:32.727510 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" event={"ID":"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1","Type":"ContainerStarted","Data":"0693e1b42c5649f9dc8f5c471ba90f4d3532951a3c7516bdfd16e7027d7e1f94"} Mar 17 15:44:34 crc kubenswrapper[4767]: I0317 15:44:34.167016 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:44:34 crc kubenswrapper[4767]: I0317 15:44:34.167548 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:44:35 crc kubenswrapper[4767]: I0317 15:44:35.747647 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" event={"ID":"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1","Type":"ContainerStarted","Data":"de9c86e43b42aff0557bac1cbc2ca7323f9ab0a0e46dff296662507f0a3679a8"} Mar 17 15:44:35 crc kubenswrapper[4767]: I0317 15:44:35.748112 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" event={"ID":"b1521838-7bb4-44a8-9ceb-758a3fbb6ae1","Type":"ContainerStarted","Data":"dff13a75c3d7941f9024ff243466027984fe16781e61cc6cdab2a7f7cc3582ef"} Mar 17 15:44:35 crc kubenswrapper[4767]: I0317 15:44:35.772313 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-zmvx6" podStartSLOduration=2.510800455 podStartE2EDuration="4.772284833s" podCreationTimestamp="2026-03-17 15:44:31 +0000 UTC" firstStartedPulling="2026-03-17 15:44:32.359710116 +0000 UTC m=+463.773026153" lastFinishedPulling="2026-03-17 15:44:34.621194484 +0000 UTC m=+466.034510531" observedRunningTime="2026-03-17 15:44:35.76755985 +0000 UTC m=+467.180875917" watchObservedRunningTime="2026-03-17 15:44:35.772284833 +0000 UTC m=+467.185600880" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.273216 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl"] Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.275812 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.279026 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.279483 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-h5c4n" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.279644 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.295662 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/70fb7838-aa67-44ed-a3bc-a54a671a90ed-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.295746 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/70fb7838-aa67-44ed-a3bc-a54a671a90ed-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.295788 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/70fb7838-aa67-44ed-a3bc-a54a671a90ed-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.295832 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvpfs\" (UniqueName: \"kubernetes.io/projected/70fb7838-aa67-44ed-a3bc-a54a671a90ed-kube-api-access-zvpfs\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.295859 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl"] Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.308830 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn"] Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.310757 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.315898 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.315898 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.316311 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-xqbnr" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.318090 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.358746 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn"] Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.400961 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/70fb7838-aa67-44ed-a3bc-a54a671a90ed-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.401041 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/70fb7838-aa67-44ed-a3bc-a54a671a90ed-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.401077 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvpfs\" (UniqueName: \"kubernetes.io/projected/70fb7838-aa67-44ed-a3bc-a54a671a90ed-kube-api-access-zvpfs\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.401142 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/70fb7838-aa67-44ed-a3bc-a54a671a90ed-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.402386 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/70fb7838-aa67-44ed-a3bc-a54a671a90ed-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: E0317 15:44:38.407496 4767 secret.go:188] Couldn't get secret openshift-monitoring/openshift-state-metrics-tls: secret "openshift-state-metrics-tls" not found Mar 17 15:44:38 crc kubenswrapper[4767]: E0317 15:44:38.407661 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70fb7838-aa67-44ed-a3bc-a54a671a90ed-openshift-state-metrics-tls podName:70fb7838-aa67-44ed-a3bc-a54a671a90ed nodeName:}" failed. No retries permitted until 2026-03-17 15:44:38.907621748 +0000 UTC m=+470.320937785 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openshift-state-metrics-tls" (UniqueName: "kubernetes.io/secret/70fb7838-aa67-44ed-a3bc-a54a671a90ed-openshift-state-metrics-tls") pod "openshift-state-metrics-566fddb674-8rdtl" (UID: "70fb7838-aa67-44ed-a3bc-a54a671a90ed") : secret "openshift-state-metrics-tls" not found Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.422908 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/70fb7838-aa67-44ed-a3bc-a54a671a90ed-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.440729 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvpfs\" (UniqueName: \"kubernetes.io/projected/70fb7838-aa67-44ed-a3bc-a54a671a90ed-kube-api-access-zvpfs\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.467615 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-l2pqv"] Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.469658 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.475841 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-pplrj" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.476138 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.476293 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.502641 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.502698 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2e0b2438-b05f-499f-9a79-e85ac5858a6b-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.502731 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml7kb\" (UniqueName: \"kubernetes.io/projected/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-api-access-ml7kb\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.502797 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.502861 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2e0b2438-b05f-499f-9a79-e85ac5858a6b-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.502924 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.604395 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.604488 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.604630 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2e0b2438-b05f-499f-9a79-e85ac5858a6b-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.604713 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lv8c\" (UniqueName: \"kubernetes.io/projected/fc1ca255-eef8-4c9f-aac8-18ba01afe527-kube-api-access-2lv8c\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.604762 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml7kb\" (UniqueName: \"kubernetes.io/projected/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-api-access-ml7kb\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.604986 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/fc1ca255-eef8-4c9f-aac8-18ba01afe527-root\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.605069 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.605153 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-wtmp\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.605306 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc1ca255-eef8-4c9f-aac8-18ba01afe527-sys\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.605394 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2e0b2438-b05f-499f-9a79-e85ac5858a6b-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.605415 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-textfile\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.605444 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc1ca255-eef8-4c9f-aac8-18ba01afe527-metrics-client-ca\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.605529 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.605649 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-tls\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.605859 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2e0b2438-b05f-499f-9a79-e85ac5858a6b-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.606275 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.606375 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2e0b2438-b05f-499f-9a79-e85ac5858a6b-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.609907 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.624580 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.626345 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml7kb\" (UniqueName: \"kubernetes.io/projected/2e0b2438-b05f-499f-9a79-e85ac5858a6b-kube-api-access-ml7kb\") pod \"kube-state-metrics-777cb5bd5d-pqrzn\" (UID: \"2e0b2438-b05f-499f-9a79-e85ac5858a6b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.659033 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.707616 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-tls\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.707714 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.707797 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lv8c\" (UniqueName: \"kubernetes.io/projected/fc1ca255-eef8-4c9f-aac8-18ba01afe527-kube-api-access-2lv8c\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.707833 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/fc1ca255-eef8-4c9f-aac8-18ba01afe527-root\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.707866 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-wtmp\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.707903 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc1ca255-eef8-4c9f-aac8-18ba01afe527-sys\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.707924 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-textfile\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.708000 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc1ca255-eef8-4c9f-aac8-18ba01afe527-metrics-client-ca\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.708066 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/fc1ca255-eef8-4c9f-aac8-18ba01afe527-root\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.708132 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc1ca255-eef8-4c9f-aac8-18ba01afe527-sys\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.708198 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-wtmp\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.709007 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-textfile\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.709305 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc1ca255-eef8-4c9f-aac8-18ba01afe527-metrics-client-ca\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.716903 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-tls\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.716894 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc1ca255-eef8-4c9f-aac8-18ba01afe527-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.733935 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lv8c\" (UniqueName: \"kubernetes.io/projected/fc1ca255-eef8-4c9f-aac8-18ba01afe527-kube-api-access-2lv8c\") pod \"node-exporter-l2pqv\" (UID: \"fc1ca255-eef8-4c9f-aac8-18ba01afe527\") " pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.792840 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-l2pqv" Mar 17 15:44:38 crc kubenswrapper[4767]: W0317 15:44:38.862851 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc1ca255_eef8_4c9f_aac8_18ba01afe527.slice/crio-f40bf9049e7adacc4c5132219a4a0128322cb2f834c229cdb1e33bb2efa210d1 WatchSource:0}: Error finding container f40bf9049e7adacc4c5132219a4a0128322cb2f834c229cdb1e33bb2efa210d1: Status 404 returned error can't find the container with id f40bf9049e7adacc4c5132219a4a0128322cb2f834c229cdb1e33bb2efa210d1 Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.920756 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/70fb7838-aa67-44ed-a3bc-a54a671a90ed-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:38 crc kubenswrapper[4767]: I0317 15:44:38.931088 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/70fb7838-aa67-44ed-a3bc-a54a671a90ed-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-8rdtl\" (UID: \"70fb7838-aa67-44ed-a3bc-a54a671a90ed\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.201024 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.202657 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn"] Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.306292 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.312213 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.334619 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.362492 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.362845 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-v2wn2" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.363049 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.367030 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.367331 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.368464 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.368581 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.369251 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.369955 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.435748 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-tls-assets\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.435849 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.435909 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.435930 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-config-volume\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.435981 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.436025 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tqd6\" (UniqueName: \"kubernetes.io/projected/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-kube-api-access-6tqd6\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.436088 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.439045 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-web-config\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.439083 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.439120 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.439222 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-config-out\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.439264 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.542215 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.542305 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-config-volume\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.542346 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.542412 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tqd6\" (UniqueName: \"kubernetes.io/projected/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-kube-api-access-6tqd6\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.542466 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.543232 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-web-config\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.543677 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.544638 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.544801 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.545094 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.545216 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.545291 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-config-out\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.545312 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.545386 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-tls-assets\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.545477 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.554739 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-web-config\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.556655 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.559807 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.560087 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.563110 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-config-volume\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.563211 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-config-out\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.567897 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tqd6\" (UniqueName: \"kubernetes.io/projected/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-kube-api-access-6tqd6\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.569260 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-tls-assets\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.582220 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/33cc0bc0-90fd-4bbd-8ae7-ab993a56232f-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f\") " pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.696707 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.770567 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl"] Mar 17 15:44:39 crc kubenswrapper[4767]: W0317 15:44:39.776040 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70fb7838_aa67_44ed_a3bc_a54a671a90ed.slice/crio-4367b0dc5bf05808832926f31b5540484272c0362f10387f00d9d846f433be35 WatchSource:0}: Error finding container 4367b0dc5bf05808832926f31b5540484272c0362f10387f00d9d846f433be35: Status 404 returned error can't find the container with id 4367b0dc5bf05808832926f31b5540484272c0362f10387f00d9d846f433be35 Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.788875 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-l2pqv" event={"ID":"fc1ca255-eef8-4c9f-aac8-18ba01afe527","Type":"ContainerStarted","Data":"f40bf9049e7adacc4c5132219a4a0128322cb2f834c229cdb1e33bb2efa210d1"} Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.797200 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" event={"ID":"2e0b2438-b05f-499f-9a79-e85ac5858a6b","Type":"ContainerStarted","Data":"4838f6c211e7f7aa69b6bb852afd1240dcd49ddb26441bc518d1d573be07b3fb"} Mar 17 15:44:39 crc kubenswrapper[4767]: I0317 15:44:39.798862 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" event={"ID":"70fb7838-aa67-44ed-a3bc-a54a671a90ed","Type":"ContainerStarted","Data":"4367b0dc5bf05808832926f31b5540484272c0362f10387f00d9d846f433be35"} Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.250765 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-89d788fd9-2skkr"] Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.278748 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-j7tgx" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.279901 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.283233 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.283797 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.284536 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.284748 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.292031 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-89d788fd9-2skkr"] Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.294844 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.294872 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-44p3k2og1ivtl" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.314951 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-q5k4q" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.322056 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.371145 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-grpc-tls\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.371242 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.371464 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.371542 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ab6a6b4-cb73-4701-a29c-f404c58966c4-metrics-client-ca\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.371584 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.371636 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s2hn\" (UniqueName: \"kubernetes.io/projected/8ab6a6b4-cb73-4701-a29c-f404c58966c4-kube-api-access-6s2hn\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.371660 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.371709 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-tls\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.397254 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ldbnt"] Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.474549 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ab6a6b4-cb73-4701-a29c-f404c58966c4-metrics-client-ca\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.473406 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ab6a6b4-cb73-4701-a29c-f404c58966c4-metrics-client-ca\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.475207 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.475245 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.476227 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s2hn\" (UniqueName: \"kubernetes.io/projected/8ab6a6b4-cb73-4701-a29c-f404c58966c4-kube-api-access-6s2hn\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.476535 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-tls\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.476611 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-grpc-tls\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.476667 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.476752 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.486253 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.486754 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.487288 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.488427 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-tls\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.488946 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.489222 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8ab6a6b4-cb73-4701-a29c-f404c58966c4-secret-grpc-tls\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.512414 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s2hn\" (UniqueName: \"kubernetes.io/projected/8ab6a6b4-cb73-4701-a29c-f404c58966c4-kube-api-access-6s2hn\") pod \"thanos-querier-89d788fd9-2skkr\" (UID: \"8ab6a6b4-cb73-4701-a29c-f404c58966c4\") " pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.619448 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.809990 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" event={"ID":"70fb7838-aa67-44ed-a3bc-a54a671a90ed","Type":"ContainerStarted","Data":"af687bf9ec2c4c1ccb7931edd6c942cc42c3a5b431b560f7ee06b0c3fba2e4e6"} Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.810043 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" event={"ID":"70fb7838-aa67-44ed-a3bc-a54a671a90ed","Type":"ContainerStarted","Data":"d146a0466c4be36ac9063f18576df3ca9397f810ec33705496a255285dad8969"} Mar 17 15:44:40 crc kubenswrapper[4767]: I0317 15:44:40.811898 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f","Type":"ContainerStarted","Data":"4cb637d092d62da5ccc53c89794e92b09c1a8d51eab8309b89db4aa34775646f"} Mar 17 15:44:41 crc kubenswrapper[4767]: I0317 15:44:41.684703 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-89d788fd9-2skkr"] Mar 17 15:44:41 crc kubenswrapper[4767]: W0317 15:44:41.689958 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ab6a6b4_cb73_4701_a29c_f404c58966c4.slice/crio-276fa15ea3109edbea11d3a03ff13d1e88b6ff30a0206417541516c1d4e63360 WatchSource:0}: Error finding container 276fa15ea3109edbea11d3a03ff13d1e88b6ff30a0206417541516c1d4e63360: Status 404 returned error can't find the container with id 276fa15ea3109edbea11d3a03ff13d1e88b6ff30a0206417541516c1d4e63360 Mar 17 15:44:41 crc kubenswrapper[4767]: I0317 15:44:41.819712 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-l2pqv" event={"ID":"fc1ca255-eef8-4c9f-aac8-18ba01afe527","Type":"ContainerStarted","Data":"da852a2ee7248c75262f8bb282ad300509f961e10764ff41a9a4ac1958162beb"} Mar 17 15:44:41 crc kubenswrapper[4767]: I0317 15:44:41.825691 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" event={"ID":"2e0b2438-b05f-499f-9a79-e85ac5858a6b","Type":"ContainerStarted","Data":"d92680b1272ff5d108f688271f354ad2df7d079641944e74dd8fd2f10d72f38d"} Mar 17 15:44:41 crc kubenswrapper[4767]: I0317 15:44:41.827837 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" event={"ID":"8ab6a6b4-cb73-4701-a29c-f404c58966c4","Type":"ContainerStarted","Data":"276fa15ea3109edbea11d3a03ff13d1e88b6ff30a0206417541516c1d4e63360"} Mar 17 15:44:42 crc kubenswrapper[4767]: I0317 15:44:42.868514 4767 generic.go:334] "Generic (PLEG): container finished" podID="fc1ca255-eef8-4c9f-aac8-18ba01afe527" containerID="da852a2ee7248c75262f8bb282ad300509f961e10764ff41a9a4ac1958162beb" exitCode=0 Mar 17 15:44:42 crc kubenswrapper[4767]: I0317 15:44:42.868633 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-l2pqv" event={"ID":"fc1ca255-eef8-4c9f-aac8-18ba01afe527","Type":"ContainerDied","Data":"da852a2ee7248c75262f8bb282ad300509f961e10764ff41a9a4ac1958162beb"} Mar 17 15:44:42 crc kubenswrapper[4767]: I0317 15:44:42.878195 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" event={"ID":"2e0b2438-b05f-499f-9a79-e85ac5858a6b","Type":"ContainerStarted","Data":"7df7383f61b507e065ff5df5c4b25e84cdedc1b1617337bce3ab02ecedbdfd16"} Mar 17 15:44:42 crc kubenswrapper[4767]: I0317 15:44:42.878311 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" event={"ID":"2e0b2438-b05f-499f-9a79-e85ac5858a6b","Type":"ContainerStarted","Data":"964b2c4402ae92c791fecd750bca965d45efec92f1bde4a01f62853a62c82629"} Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.288154 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-pqrzn" podStartSLOduration=3.15154594 podStartE2EDuration="5.288131895s" podCreationTimestamp="2026-03-17 15:44:38 +0000 UTC" firstStartedPulling="2026-03-17 15:44:39.225187019 +0000 UTC m=+470.638503066" lastFinishedPulling="2026-03-17 15:44:41.361772974 +0000 UTC m=+472.775089021" observedRunningTime="2026-03-17 15:44:42.959404493 +0000 UTC m=+474.372720550" watchObservedRunningTime="2026-03-17 15:44:43.288131895 +0000 UTC m=+474.701447942" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.291187 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7948d5b58d-tvgzz"] Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.292098 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.442792 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7948d5b58d-tvgzz"] Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.476967 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-serving-cert\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.477033 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-service-ca\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.477123 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-trusted-ca-bundle\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.477365 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-oauth-config\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.477413 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-oauth-serving-cert\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.477572 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzzrq\" (UniqueName: \"kubernetes.io/projected/14549ef2-a133-45bf-88b6-6ea5741c6286-kube-api-access-tzzrq\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.477591 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-console-config\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.580419 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzzrq\" (UniqueName: \"kubernetes.io/projected/14549ef2-a133-45bf-88b6-6ea5741c6286-kube-api-access-tzzrq\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.580476 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-console-config\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.580577 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-serving-cert\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.580615 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-service-ca\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.580660 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-trusted-ca-bundle\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.580724 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-oauth-config\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.580750 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-oauth-serving-cert\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.582359 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-oauth-serving-cert\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.582495 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-service-ca\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.583307 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-console-config\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.585441 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-trusted-ca-bundle\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.608042 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-oauth-config\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.608700 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-serving-cert\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.685280 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzzrq\" (UniqueName: \"kubernetes.io/projected/14549ef2-a133-45bf-88b6-6ea5741c6286-kube-api-access-tzzrq\") pod \"console-7948d5b58d-tvgzz\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.821231 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-56684ddd8d-zrdxq"] Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.822600 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.827600 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-erkp65hnd1h84" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.827860 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.827780 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.828272 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-5frrl" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.843206 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.848280 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-56684ddd8d-zrdxq"] Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.848423 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.924243 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.992455 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/5a2cfb18-0930-434f-a9bc-824575209dda-secret-metrics-client-certs\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.992546 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a2cfb18-0930-434f-a9bc-824575209dda-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.992606 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/5a2cfb18-0930-434f-a9bc-824575209dda-secret-metrics-server-tls\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.992640 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/5a2cfb18-0930-434f-a9bc-824575209dda-audit-log\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.992923 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/5a2cfb18-0930-434f-a9bc-824575209dda-metrics-server-audit-profiles\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.993124 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a2cfb18-0930-434f-a9bc-824575209dda-client-ca-bundle\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.993248 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77xg5\" (UniqueName: \"kubernetes.io/projected/5a2cfb18-0930-434f-a9bc-824575209dda-kube-api-access-77xg5\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:43 crc kubenswrapper[4767]: I0317 15:44:43.999050 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5"] Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.000462 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.002981 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.003390 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.028104 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5"] Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.119878 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/5a2cfb18-0930-434f-a9bc-824575209dda-secret-metrics-client-certs\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.119935 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a2cfb18-0930-434f-a9bc-824575209dda-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.119970 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/5a2cfb18-0930-434f-a9bc-824575209dda-secret-metrics-server-tls\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.119995 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/5a2cfb18-0930-434f-a9bc-824575209dda-audit-log\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.120039 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/5a2cfb18-0930-434f-a9bc-824575209dda-metrics-server-audit-profiles\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.120090 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a2cfb18-0930-434f-a9bc-824575209dda-client-ca-bundle\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.120107 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77xg5\" (UniqueName: \"kubernetes.io/projected/5a2cfb18-0930-434f-a9bc-824575209dda-kube-api-access-77xg5\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.121102 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/5a2cfb18-0930-434f-a9bc-824575209dda-audit-log\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.121777 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a2cfb18-0930-434f-a9bc-824575209dda-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.123326 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/5a2cfb18-0930-434f-a9bc-824575209dda-metrics-server-audit-profiles\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.131093 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a2cfb18-0930-434f-a9bc-824575209dda-client-ca-bundle\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.131239 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/5a2cfb18-0930-434f-a9bc-824575209dda-secret-metrics-client-certs\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.132773 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/5a2cfb18-0930-434f-a9bc-824575209dda-secret-metrics-server-tls\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.153727 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77xg5\" (UniqueName: \"kubernetes.io/projected/5a2cfb18-0930-434f-a9bc-824575209dda-kube-api-access-77xg5\") pod \"metrics-server-56684ddd8d-zrdxq\" (UID: \"5a2cfb18-0930-434f-a9bc-824575209dda\") " pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.223061 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/93033ac4-fd81-46eb-8014-184056ab6de2-monitoring-plugin-cert\") pod \"monitoring-plugin-6bbd5f6758-kwcw5\" (UID: \"93033ac4-fd81-46eb-8014-184056ab6de2\") " pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.501455 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.501941 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/93033ac4-fd81-46eb-8014-184056ab6de2-monitoring-plugin-cert\") pod \"monitoring-plugin-6bbd5f6758-kwcw5\" (UID: \"93033ac4-fd81-46eb-8014-184056ab6de2\") " pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.507719 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/93033ac4-fd81-46eb-8014-184056ab6de2-monitoring-plugin-cert\") pod \"monitoring-plugin-6bbd5f6758-kwcw5\" (UID: \"93033ac4-fd81-46eb-8014-184056ab6de2\") " pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" Mar 17 15:44:44 crc kubenswrapper[4767]: I0317 15:44:44.732203 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.227212 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.286780 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.558582 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.559124 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.559356 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.559645 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.559769 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.559883 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.559981 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.560087 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.560337 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-8v8vc" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.560403 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.560566 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.561662 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-5cgd53rj4cp71" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563749 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563788 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/60ee8af2-4ae0-402e-8a13-2dce36db4bab-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563810 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563833 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563849 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563856 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/60ee8af2-4ae0-402e-8a13-2dce36db4bab-config-out\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563877 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-web-config\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563894 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563923 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563937 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563966 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-config\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.563983 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.564006 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9xrm\" (UniqueName: \"kubernetes.io/projected/60ee8af2-4ae0-402e-8a13-2dce36db4bab-kube-api-access-q9xrm\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.564036 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.564068 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.564202 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.564257 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.564280 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/60ee8af2-4ae0-402e-8a13-2dce36db4bab-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.564307 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.583933 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.766880 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9xrm\" (UniqueName: \"kubernetes.io/projected/60ee8af2-4ae0-402e-8a13-2dce36db4bab-kube-api-access-q9xrm\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.766934 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.766962 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.766990 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767038 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767061 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/60ee8af2-4ae0-402e-8a13-2dce36db4bab-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767519 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767569 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767597 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/60ee8af2-4ae0-402e-8a13-2dce36db4bab-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767636 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767666 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767712 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/60ee8af2-4ae0-402e-8a13-2dce36db4bab-config-out\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767732 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-web-config\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767756 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767805 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767831 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767867 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-config\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.767893 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.768575 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/60ee8af2-4ae0-402e-8a13-2dce36db4bab-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.768853 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.771398 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.774155 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/60ee8af2-4ae0-402e-8a13-2dce36db4bab-config-out\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.775788 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.777287 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.783136 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.783467 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/60ee8af2-4ae0-402e-8a13-2dce36db4bab-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.787657 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.795889 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-web-config\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.797249 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.798826 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.798932 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/60ee8af2-4ae0-402e-8a13-2dce36db4bab-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.805412 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-config\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.806233 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9xrm\" (UniqueName: \"kubernetes.io/projected/60ee8af2-4ae0-402e-8a13-2dce36db4bab-kube-api-access-q9xrm\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.807165 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.808258 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.809789 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/60ee8af2-4ae0-402e-8a13-2dce36db4bab-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"60ee8af2-4ae0-402e-8a13-2dce36db4bab\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:45 crc kubenswrapper[4767]: I0317 15:44:45.879256 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:44:46 crc kubenswrapper[4767]: I0317 15:44:46.088694 4767 generic.go:334] "Generic (PLEG): container finished" podID="33cc0bc0-90fd-4bbd-8ae7-ab993a56232f" containerID="73c32311ab01cb7cdf4ff127b8f35b4af3534e1bcddb1b2a26f1e7bfa655fe90" exitCode=0 Mar 17 15:44:46 crc kubenswrapper[4767]: I0317 15:44:46.088784 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f","Type":"ContainerDied","Data":"73c32311ab01cb7cdf4ff127b8f35b4af3534e1bcddb1b2a26f1e7bfa655fe90"} Mar 17 15:44:46 crc kubenswrapper[4767]: I0317 15:44:46.103303 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-l2pqv" event={"ID":"fc1ca255-eef8-4c9f-aac8-18ba01afe527","Type":"ContainerStarted","Data":"ba64b0409f73cfca4dec9ce7fea890f96c15ab19c430a82a9637a3b8c1d70f3d"} Mar 17 15:44:46 crc kubenswrapper[4767]: I0317 15:44:46.639769 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7948d5b58d-tvgzz"] Mar 17 15:44:46 crc kubenswrapper[4767]: I0317 15:44:46.733070 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5"] Mar 17 15:44:46 crc kubenswrapper[4767]: I0317 15:44:46.741663 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-56684ddd8d-zrdxq"] Mar 17 15:44:46 crc kubenswrapper[4767]: I0317 15:44:46.753505 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Mar 17 15:44:46 crc kubenswrapper[4767]: W0317 15:44:46.769428 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60ee8af2_4ae0_402e_8a13_2dce36db4bab.slice/crio-ae7ec88b868cb09b4cee0d56f6537df8720abd2ad03fd713aae4c90033770738 WatchSource:0}: Error finding container ae7ec88b868cb09b4cee0d56f6537df8720abd2ad03fd713aae4c90033770738: Status 404 returned error can't find the container with id ae7ec88b868cb09b4cee0d56f6537df8720abd2ad03fd713aae4c90033770738 Mar 17 15:44:46 crc kubenswrapper[4767]: W0317 15:44:46.773361 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a2cfb18_0930_434f_a9bc_824575209dda.slice/crio-8eaabea61987406bf8f6ea8b48edac79cfe158c2894721f637d65eceb20e0d5c WatchSource:0}: Error finding container 8eaabea61987406bf8f6ea8b48edac79cfe158c2894721f637d65eceb20e0d5c: Status 404 returned error can't find the container with id 8eaabea61987406bf8f6ea8b48edac79cfe158c2894721f637d65eceb20e0d5c Mar 17 15:44:47 crc kubenswrapper[4767]: I0317 15:44:47.124873 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" event={"ID":"5a2cfb18-0930-434f-a9bc-824575209dda","Type":"ContainerStarted","Data":"8eaabea61987406bf8f6ea8b48edac79cfe158c2894721f637d65eceb20e0d5c"} Mar 17 15:44:47 crc kubenswrapper[4767]: I0317 15:44:47.127052 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"60ee8af2-4ae0-402e-8a13-2dce36db4bab","Type":"ContainerStarted","Data":"c2f04fb3449f26afa19ea74c993c166ed78a983a5b312919cd05b23503e626d0"} Mar 17 15:44:47 crc kubenswrapper[4767]: I0317 15:44:47.127104 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"60ee8af2-4ae0-402e-8a13-2dce36db4bab","Type":"ContainerStarted","Data":"ae7ec88b868cb09b4cee0d56f6537df8720abd2ad03fd713aae4c90033770738"} Mar 17 15:44:47 crc kubenswrapper[4767]: I0317 15:44:47.130698 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7948d5b58d-tvgzz" event={"ID":"14549ef2-a133-45bf-88b6-6ea5741c6286","Type":"ContainerStarted","Data":"7c4221f87ad6df0f0a53c503ef6526d4ab75fb0400000c8ef67df57fc592a371"} Mar 17 15:44:47 crc kubenswrapper[4767]: I0317 15:44:47.130755 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7948d5b58d-tvgzz" event={"ID":"14549ef2-a133-45bf-88b6-6ea5741c6286","Type":"ContainerStarted","Data":"974e083983451ce4d9d44782a7454363b1f66b66f4ae1037b30d038ade9d7e56"} Mar 17 15:44:47 crc kubenswrapper[4767]: I0317 15:44:47.137411 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-l2pqv" event={"ID":"fc1ca255-eef8-4c9f-aac8-18ba01afe527","Type":"ContainerStarted","Data":"b35f57e5c8e1aca1338524d4adca8017f01e474b4d4c7132eae1f86997eb9965"} Mar 17 15:44:47 crc kubenswrapper[4767]: I0317 15:44:47.142981 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" event={"ID":"93033ac4-fd81-46eb-8014-184056ab6de2","Type":"ContainerStarted","Data":"55198eeb4ba73807a16b48b430afcf883fb13965c25395dc25b775246316b661"} Mar 17 15:44:47 crc kubenswrapper[4767]: I0317 15:44:47.254730 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-l2pqv" podStartSLOduration=6.753218998 podStartE2EDuration="9.254704037s" podCreationTimestamp="2026-03-17 15:44:38 +0000 UTC" firstStartedPulling="2026-03-17 15:44:38.866163054 +0000 UTC m=+470.279479101" lastFinishedPulling="2026-03-17 15:44:41.367648093 +0000 UTC m=+472.780964140" observedRunningTime="2026-03-17 15:44:47.253812409 +0000 UTC m=+478.667128456" watchObservedRunningTime="2026-03-17 15:44:47.254704037 +0000 UTC m=+478.668020084" Mar 17 15:44:47 crc kubenswrapper[4767]: I0317 15:44:47.280953 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7948d5b58d-tvgzz" podStartSLOduration=4.280924705 podStartE2EDuration="4.280924705s" podCreationTimestamp="2026-03-17 15:44:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:44:47.2774578 +0000 UTC m=+478.690773867" watchObservedRunningTime="2026-03-17 15:44:47.280924705 +0000 UTC m=+478.694240752" Mar 17 15:44:48 crc kubenswrapper[4767]: I0317 15:44:48.159517 4767 generic.go:334] "Generic (PLEG): container finished" podID="60ee8af2-4ae0-402e-8a13-2dce36db4bab" containerID="c2f04fb3449f26afa19ea74c993c166ed78a983a5b312919cd05b23503e626d0" exitCode=0 Mar 17 15:44:48 crc kubenswrapper[4767]: I0317 15:44:48.159677 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"60ee8af2-4ae0-402e-8a13-2dce36db4bab","Type":"ContainerDied","Data":"c2f04fb3449f26afa19ea74c993c166ed78a983a5b312919cd05b23503e626d0"} Mar 17 15:44:48 crc kubenswrapper[4767]: I0317 15:44:48.170532 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" event={"ID":"70fb7838-aa67-44ed-a3bc-a54a671a90ed","Type":"ContainerStarted","Data":"0b3f562127bb117f3395e68695111ab670c6c6ccdc3ab12aa2af5607b9c1c4e0"} Mar 17 15:44:49 crc kubenswrapper[4767]: I0317 15:44:49.411522 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-8rdtl" podStartSLOduration=3.79933254 podStartE2EDuration="11.411492576s" podCreationTimestamp="2026-03-17 15:44:38 +0000 UTC" firstStartedPulling="2026-03-17 15:44:40.290936639 +0000 UTC m=+471.704252686" lastFinishedPulling="2026-03-17 15:44:47.903096675 +0000 UTC m=+479.316412722" observedRunningTime="2026-03-17 15:44:48.271195076 +0000 UTC m=+479.684511143" watchObservedRunningTime="2026-03-17 15:44:49.411492576 +0000 UTC m=+480.824808623" Mar 17 15:44:53 crc kubenswrapper[4767]: I0317 15:44:53.925674 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:53 crc kubenswrapper[4767]: I0317 15:44:53.929483 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:53 crc kubenswrapper[4767]: I0317 15:44:53.940985 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:54 crc kubenswrapper[4767]: I0317 15:44:54.389447 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:44:54 crc kubenswrapper[4767]: I0317 15:44:54.755197 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-9cz8q"] Mar 17 15:44:57 crc kubenswrapper[4767]: I0317 15:44:57.432939 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" event={"ID":"8ab6a6b4-cb73-4701-a29c-f404c58966c4","Type":"ContainerStarted","Data":"42642521634d4fcc74a2496d7f2eb25fcc67792d8a0171aa407c4b8d2c26c75f"} Mar 17 15:44:57 crc kubenswrapper[4767]: I0317 15:44:57.439141 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f","Type":"ContainerStarted","Data":"92f5f80c487cb25a13880d38d5fbaf4d66e4e3d0565cdeeaa2c0e4266766aa7d"} Mar 17 15:44:57 crc kubenswrapper[4767]: I0317 15:44:57.441949 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" event={"ID":"93033ac4-fd81-46eb-8014-184056ab6de2","Type":"ContainerStarted","Data":"17eea40e328b14d4012f63ea3eb349336b5b70dd7780e706f5fe5b395e926ad0"} Mar 17 15:44:57 crc kubenswrapper[4767]: I0317 15:44:57.442793 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" Mar 17 15:44:57 crc kubenswrapper[4767]: I0317 15:44:57.445196 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" event={"ID":"5a2cfb18-0930-434f-a9bc-824575209dda","Type":"ContainerStarted","Data":"4b9cbea47b435d9e0e78f8631f1543a80411175fc2fb9795a340ded81d25b8ca"} Mar 17 15:44:57 crc kubenswrapper[4767]: I0317 15:44:57.622679 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" Mar 17 15:44:57 crc kubenswrapper[4767]: I0317 15:44:57.626319 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"60ee8af2-4ae0-402e-8a13-2dce36db4bab","Type":"ContainerStarted","Data":"3beae19ec1ada40e47887af86ae4f6e5d467410db06ca28d05af86700e88d0fc"} Mar 17 15:44:57 crc kubenswrapper[4767]: I0317 15:44:57.653015 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" podStartSLOduration=4.3977964 podStartE2EDuration="14.652959729s" podCreationTimestamp="2026-03-17 15:44:43 +0000 UTC" firstStartedPulling="2026-03-17 15:44:46.752659971 +0000 UTC m=+478.165976018" lastFinishedPulling="2026-03-17 15:44:57.0078233 +0000 UTC m=+488.421139347" observedRunningTime="2026-03-17 15:44:57.644468111 +0000 UTC m=+489.057784158" watchObservedRunningTime="2026-03-17 15:44:57.652959729 +0000 UTC m=+489.066275776" Mar 17 15:44:57 crc kubenswrapper[4767]: I0317 15:44:57.694754 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" podStartSLOduration=4.575031383 podStartE2EDuration="14.694728171s" podCreationTimestamp="2026-03-17 15:44:43 +0000 UTC" firstStartedPulling="2026-03-17 15:44:46.886700388 +0000 UTC m=+478.300016435" lastFinishedPulling="2026-03-17 15:44:57.006397176 +0000 UTC m=+488.419713223" observedRunningTime="2026-03-17 15:44:57.692471913 +0000 UTC m=+489.105787980" watchObservedRunningTime="2026-03-17 15:44:57.694728171 +0000 UTC m=+489.108044218" Mar 17 15:44:58 crc kubenswrapper[4767]: I0317 15:44:58.644046 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"60ee8af2-4ae0-402e-8a13-2dce36db4bab","Type":"ContainerStarted","Data":"d3ee03c083dfd317927907bd313e0870d83400bb35ea76b79a1db44cb3307af1"} Mar 17 15:44:58 crc kubenswrapper[4767]: I0317 15:44:58.648554 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" event={"ID":"8ab6a6b4-cb73-4701-a29c-f404c58966c4","Type":"ContainerStarted","Data":"4a3723518defae9ab08adaf70e1304b7c8d4a6204a903a8b47bfc68aff5bc6e5"} Mar 17 15:44:58 crc kubenswrapper[4767]: I0317 15:44:58.653646 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f","Type":"ContainerStarted","Data":"ec9e95c30c8c5b579745178910e95f36867dd261f4e3a2ef70c09bb8f21aab0c"} Mar 17 15:44:59 crc kubenswrapper[4767]: I0317 15:44:59.732716 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" event={"ID":"8ab6a6b4-cb73-4701-a29c-f404c58966c4","Type":"ContainerStarted","Data":"4f9e0ec5ec8c77e1a2e27b88bc7a046ad6b7c4677649157552020b192c893b18"} Mar 17 15:44:59 crc kubenswrapper[4767]: I0317 15:44:59.738966 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f","Type":"ContainerStarted","Data":"81c578cd3fad635b97042652d264aace2b33ee64bedaca7a59e7ff0ac32d2960"} Mar 17 15:44:59 crc kubenswrapper[4767]: I0317 15:44:59.739040 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f","Type":"ContainerStarted","Data":"ff391d4d60e3949b1a7e84cb0e461d3394284d624350d4331ae98094f6338f2c"} Mar 17 15:44:59 crc kubenswrapper[4767]: I0317 15:44:59.739076 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f","Type":"ContainerStarted","Data":"6690da6edb44a53ccc9688dd324b226cea13abae2339de84e8ff07c46ebd8161"} Mar 17 15:44:59 crc kubenswrapper[4767]: I0317 15:44:59.745455 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"60ee8af2-4ae0-402e-8a13-2dce36db4bab","Type":"ContainerStarted","Data":"36626c01d58fb05fc1784daa29db5959fc3ee22d6b723ba04b318b4805e08a96"} Mar 17 15:44:59 crc kubenswrapper[4767]: I0317 15:44:59.745491 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"60ee8af2-4ae0-402e-8a13-2dce36db4bab","Type":"ContainerStarted","Data":"b358cb16adcf58419c23a7f0a22ca0c87a155a08cbb6483e5a9df9dce2d2f45f"} Mar 17 15:44:59 crc kubenswrapper[4767]: I0317 15:44:59.745504 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"60ee8af2-4ae0-402e-8a13-2dce36db4bab","Type":"ContainerStarted","Data":"e8cec7d501e1a2641a87d1946e5ff92fb9bc74457c7c64da3bc2072c46c64d21"} Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.156518 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562705-775td"] Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.157859 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.161135 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.162566 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.177966 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562705-775td"] Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.287726 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7fcd\" (UniqueName: \"kubernetes.io/projected/dd451935-a4a7-41df-aa19-8e7065df2b0c-kube-api-access-t7fcd\") pod \"collect-profiles-29562705-775td\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.287912 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd451935-a4a7-41df-aa19-8e7065df2b0c-secret-volume\") pod \"collect-profiles-29562705-775td\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.287979 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd451935-a4a7-41df-aa19-8e7065df2b0c-config-volume\") pod \"collect-profiles-29562705-775td\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.389451 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd451935-a4a7-41df-aa19-8e7065df2b0c-secret-volume\") pod \"collect-profiles-29562705-775td\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.389523 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd451935-a4a7-41df-aa19-8e7065df2b0c-config-volume\") pod \"collect-profiles-29562705-775td\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.389647 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7fcd\" (UniqueName: \"kubernetes.io/projected/dd451935-a4a7-41df-aa19-8e7065df2b0c-kube-api-access-t7fcd\") pod \"collect-profiles-29562705-775td\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.391795 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd451935-a4a7-41df-aa19-8e7065df2b0c-config-volume\") pod \"collect-profiles-29562705-775td\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.403057 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd451935-a4a7-41df-aa19-8e7065df2b0c-secret-volume\") pod \"collect-profiles-29562705-775td\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.409407 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7fcd\" (UniqueName: \"kubernetes.io/projected/dd451935-a4a7-41df-aa19-8e7065df2b0c-kube-api-access-t7fcd\") pod \"collect-profiles-29562705-775td\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:00 crc kubenswrapper[4767]: I0317 15:45:00.500563 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:01 crc kubenswrapper[4767]: I0317 15:45:01.698034 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"60ee8af2-4ae0-402e-8a13-2dce36db4bab","Type":"ContainerStarted","Data":"2716c4c3a22dc06e3f1c252aead4c28d24b809a967bbc70d04069e20d9f49d2b"} Mar 17 15:45:01 crc kubenswrapper[4767]: I0317 15:45:01.734875 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=7.8110424389999995 podStartE2EDuration="16.734852833s" podCreationTimestamp="2026-03-17 15:44:45 +0000 UTC" firstStartedPulling="2026-03-17 15:44:48.164206528 +0000 UTC m=+479.577522585" lastFinishedPulling="2026-03-17 15:44:57.088016932 +0000 UTC m=+488.501332979" observedRunningTime="2026-03-17 15:45:01.7341064 +0000 UTC m=+493.147422467" watchObservedRunningTime="2026-03-17 15:45:01.734852833 +0000 UTC m=+493.148168880" Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.433518 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562705-775td"] Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.711062 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" event={"ID":"dd451935-a4a7-41df-aa19-8e7065df2b0c","Type":"ContainerStarted","Data":"1b4446556f867f7b88071dbfee960e4840070cf0d82d6628ef03d011f3a727bc"} Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.711822 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" event={"ID":"dd451935-a4a7-41df-aa19-8e7065df2b0c","Type":"ContainerStarted","Data":"945421619e3fb8ccdd24b567c9e5bebe37feb97342721f81132554b6646cf32a"} Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.717153 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" event={"ID":"8ab6a6b4-cb73-4701-a29c-f404c58966c4","Type":"ContainerStarted","Data":"9d712b6cfa687408fdaa09e2f50d8e99953b6cb98ace23f8cfca56673ca534bd"} Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.717272 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" event={"ID":"8ab6a6b4-cb73-4701-a29c-f404c58966c4","Type":"ContainerStarted","Data":"a2a35be736ebb2a283610db765e18795d11b3dee9ea7f2e5b210968020bd43a7"} Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.717296 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" event={"ID":"8ab6a6b4-cb73-4701-a29c-f404c58966c4","Type":"ContainerStarted","Data":"63040bad772ba64378d042b8c9454351117cdc40a1443a439891af041982ef86"} Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.717335 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.725668 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"33cc0bc0-90fd-4bbd-8ae7-ab993a56232f","Type":"ContainerStarted","Data":"8a0969ed13b98dabb13ffaff8dbec684300368279190e924035dec7eca4dd6b7"} Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.742575 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" podStartSLOduration=2.742536275 podStartE2EDuration="2.742536275s" podCreationTimestamp="2026-03-17 15:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:45:02.73547398 +0000 UTC m=+494.148790027" watchObservedRunningTime="2026-03-17 15:45:02.742536275 +0000 UTC m=+494.155852322" Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.762117 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.837796 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.33471926 podStartE2EDuration="23.837762455s" podCreationTimestamp="2026-03-17 15:44:39 +0000 UTC" firstStartedPulling="2026-03-17 15:44:40.33038466 +0000 UTC m=+471.743700707" lastFinishedPulling="2026-03-17 15:45:01.833427855 +0000 UTC m=+493.246743902" observedRunningTime="2026-03-17 15:45:02.827911335 +0000 UTC m=+494.241227412" watchObservedRunningTime="2026-03-17 15:45:02.837762455 +0000 UTC m=+494.251078512" Mar 17 15:45:02 crc kubenswrapper[4767]: I0317 15:45:02.937270 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" podStartSLOduration=2.796616564 podStartE2EDuration="22.937241454s" podCreationTimestamp="2026-03-17 15:44:40 +0000 UTC" firstStartedPulling="2026-03-17 15:44:41.692800895 +0000 UTC m=+473.106116942" lastFinishedPulling="2026-03-17 15:45:01.833425785 +0000 UTC m=+493.246741832" observedRunningTime="2026-03-17 15:45:02.926798896 +0000 UTC m=+494.340114963" watchObservedRunningTime="2026-03-17 15:45:02.937241454 +0000 UTC m=+494.350557491" Mar 17 15:45:03 crc kubenswrapper[4767]: I0317 15:45:03.741085 4767 generic.go:334] "Generic (PLEG): container finished" podID="dd451935-a4a7-41df-aa19-8e7065df2b0c" containerID="1b4446556f867f7b88071dbfee960e4840070cf0d82d6628ef03d011f3a727bc" exitCode=0 Mar 17 15:45:03 crc kubenswrapper[4767]: I0317 15:45:03.741159 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" event={"ID":"dd451935-a4a7-41df-aa19-8e7065df2b0c","Type":"ContainerDied","Data":"1b4446556f867f7b88071dbfee960e4840070cf0d82d6628ef03d011f3a727bc"} Mar 17 15:45:04 crc kubenswrapper[4767]: I0317 15:45:04.381475 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:45:04 crc kubenswrapper[4767]: I0317 15:45:04.388601 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:45:04 crc kubenswrapper[4767]: I0317 15:45:04.388743 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:45:04 crc kubenswrapper[4767]: I0317 15:45:04.389989 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c904b546ac42362d6e52f6a82d664f6ec63ec5a64d0aa1a9ae36b3788164828"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 15:45:04 crc kubenswrapper[4767]: I0317 15:45:04.390074 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://3c904b546ac42362d6e52f6a82d664f6ec63ec5a64d0aa1a9ae36b3788164828" gracePeriod=600 Mar 17 15:45:04 crc kubenswrapper[4767]: I0317 15:45:04.503451 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:45:04 crc kubenswrapper[4767]: I0317 15:45:04.504490 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:45:04 crc kubenswrapper[4767]: I0317 15:45:04.802551 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="3c904b546ac42362d6e52f6a82d664f6ec63ec5a64d0aa1a9ae36b3788164828" exitCode=0 Mar 17 15:45:04 crc kubenswrapper[4767]: I0317 15:45:04.804297 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"3c904b546ac42362d6e52f6a82d664f6ec63ec5a64d0aa1a9ae36b3788164828"} Mar 17 15:45:04 crc kubenswrapper[4767]: I0317 15:45:04.804563 4767 scope.go:117] "RemoveContainer" containerID="9d51c5534df2e3ac93f8cbe502e4183e5d0c79328004ef9141c9c0c12727983b" Mar 17 15:45:05 crc kubenswrapper[4767]: I0317 15:45:05.446343 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" podUID="83518d71-dc9f-4e6c-9ad4-66ee45c262f8" containerName="registry" containerID="cri-o://3e2487e816fde5f768d7d7f328bda10584be833b3a25b5b7497b053e59a49751" gracePeriod=30 Mar 17 15:45:05 crc kubenswrapper[4767]: I0317 15:45:05.590087 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.086537 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd451935-a4a7-41df-aa19-8e7065df2b0c-secret-volume\") pod \"dd451935-a4a7-41df-aa19-8e7065df2b0c\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.086890 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7fcd\" (UniqueName: \"kubernetes.io/projected/dd451935-a4a7-41df-aa19-8e7065df2b0c-kube-api-access-t7fcd\") pod \"dd451935-a4a7-41df-aa19-8e7065df2b0c\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.087051 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd451935-a4a7-41df-aa19-8e7065df2b0c-config-volume\") pod \"dd451935-a4a7-41df-aa19-8e7065df2b0c\" (UID: \"dd451935-a4a7-41df-aa19-8e7065df2b0c\") " Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.090668 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd451935-a4a7-41df-aa19-8e7065df2b0c-config-volume" (OuterVolumeSpecName: "config-volume") pod "dd451935-a4a7-41df-aa19-8e7065df2b0c" (UID: "dd451935-a4a7-41df-aa19-8e7065df2b0c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.091942 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.102058 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd451935-a4a7-41df-aa19-8e7065df2b0c-kube-api-access-t7fcd" (OuterVolumeSpecName: "kube-api-access-t7fcd") pod "dd451935-a4a7-41df-aa19-8e7065df2b0c" (UID: "dd451935-a4a7-41df-aa19-8e7065df2b0c"). InnerVolumeSpecName "kube-api-access-t7fcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.113756 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd451935-a4a7-41df-aa19-8e7065df2b0c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dd451935-a4a7-41df-aa19-8e7065df2b0c" (UID: "dd451935-a4a7-41df-aa19-8e7065df2b0c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.133079 4767 generic.go:334] "Generic (PLEG): container finished" podID="83518d71-dc9f-4e6c-9ad4-66ee45c262f8" containerID="3e2487e816fde5f768d7d7f328bda10584be833b3a25b5b7497b053e59a49751" exitCode=0 Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.133217 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" event={"ID":"83518d71-dc9f-4e6c-9ad4-66ee45c262f8","Type":"ContainerDied","Data":"3e2487e816fde5f768d7d7f328bda10584be833b3a25b5b7497b053e59a49751"} Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.141667 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.142778 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562705-775td" event={"ID":"dd451935-a4a7-41df-aa19-8e7065df2b0c","Type":"ContainerDied","Data":"945421619e3fb8ccdd24b567c9e5bebe37feb97342721f81132554b6646cf32a"} Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.142878 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="945421619e3fb8ccdd24b567c9e5bebe37feb97342721f81132554b6646cf32a" Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.149766 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"d152bc4b2f69796edf7aec43d7f21284777821024a613d49605f7c9d9db18719"} Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.189541 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7fcd\" (UniqueName: \"kubernetes.io/projected/dd451935-a4a7-41df-aa19-8e7065df2b0c-kube-api-access-t7fcd\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.190189 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd451935-a4a7-41df-aa19-8e7065df2b0c-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.190821 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd451935-a4a7-41df-aa19-8e7065df2b0c-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:06 crc kubenswrapper[4767]: I0317 15:45:06.728637 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.052148 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-trusted-ca\") pod \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.052261 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-installation-pull-secrets\") pod \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.052320 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-tls\") pod \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.052347 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-certificates\") pod \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.052822 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.052952 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-ca-trust-extracted\") pod \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.052984 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-bound-sa-token\") pod \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.053010 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzmxx\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-kube-api-access-pzmxx\") pod \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\" (UID: \"83518d71-dc9f-4e6c-9ad4-66ee45c262f8\") " Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.054197 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "83518d71-dc9f-4e6c-9ad4-66ee45c262f8" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.054274 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "83518d71-dc9f-4e6c-9ad4-66ee45c262f8" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.058620 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-kube-api-access-pzmxx" (OuterVolumeSpecName: "kube-api-access-pzmxx") pod "83518d71-dc9f-4e6c-9ad4-66ee45c262f8" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8"). InnerVolumeSpecName "kube-api-access-pzmxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.060105 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "83518d71-dc9f-4e6c-9ad4-66ee45c262f8" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.068293 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "83518d71-dc9f-4e6c-9ad4-66ee45c262f8" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.068610 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "83518d71-dc9f-4e6c-9ad4-66ee45c262f8" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.075678 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "83518d71-dc9f-4e6c-9ad4-66ee45c262f8" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.080574 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "83518d71-dc9f-4e6c-9ad4-66ee45c262f8" (UID: "83518d71-dc9f-4e6c-9ad4-66ee45c262f8"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.155475 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.155531 4767 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.155546 4767 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.155564 4767 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.155575 4767 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.155638 4767 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.155651 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzmxx\" (UniqueName: \"kubernetes.io/projected/83518d71-dc9f-4e6c-9ad4-66ee45c262f8-kube-api-access-pzmxx\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.162247 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" event={"ID":"83518d71-dc9f-4e6c-9ad4-66ee45c262f8","Type":"ContainerDied","Data":"74361c254ab87b351356f4fee0255dc0c45e89d710e2fdddefc488a754694d35"} Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.162353 4767 scope.go:117] "RemoveContainer" containerID="3e2487e816fde5f768d7d7f328bda10584be833b3a25b5b7497b053e59a49751" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.162391 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ldbnt" Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.206432 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ldbnt"] Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.211847 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ldbnt"] Mar 17 15:45:07 crc kubenswrapper[4767]: I0317 15:45:07.367828 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83518d71-dc9f-4e6c-9ad4-66ee45c262f8" path="/var/lib/kubelet/pods/83518d71-dc9f-4e6c-9ad4-66ee45c262f8/volumes" Mar 17 15:45:19 crc kubenswrapper[4767]: I0317 15:45:19.807670 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-9cz8q" podUID="deba5665-b026-4d87-a986-b17f02f644d7" containerName="console" containerID="cri-o://415bc454f7b25622d72c606b86b13b8ba7e778cd5cef1a1a64a9fb92f886fbf6" gracePeriod=15 Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.053531 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-9cz8q_deba5665-b026-4d87-a986-b17f02f644d7/console/0.log" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.053998 4767 generic.go:334] "Generic (PLEG): container finished" podID="deba5665-b026-4d87-a986-b17f02f644d7" containerID="415bc454f7b25622d72c606b86b13b8ba7e778cd5cef1a1a64a9fb92f886fbf6" exitCode=2 Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.054041 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9cz8q" event={"ID":"deba5665-b026-4d87-a986-b17f02f644d7","Type":"ContainerDied","Data":"415bc454f7b25622d72c606b86b13b8ba7e778cd5cef1a1a64a9fb92f886fbf6"} Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.301969 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-9cz8q_deba5665-b026-4d87-a986-b17f02f644d7/console/0.log" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.302073 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.337554 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-serving-cert\") pod \"deba5665-b026-4d87-a986-b17f02f644d7\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.337689 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-oauth-config\") pod \"deba5665-b026-4d87-a986-b17f02f644d7\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.337735 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-trusted-ca-bundle\") pod \"deba5665-b026-4d87-a986-b17f02f644d7\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.337785 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-console-config\") pod \"deba5665-b026-4d87-a986-b17f02f644d7\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.337832 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-oauth-serving-cert\") pod \"deba5665-b026-4d87-a986-b17f02f644d7\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.337921 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls8gg\" (UniqueName: \"kubernetes.io/projected/deba5665-b026-4d87-a986-b17f02f644d7-kube-api-access-ls8gg\") pod \"deba5665-b026-4d87-a986-b17f02f644d7\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.337943 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-service-ca\") pod \"deba5665-b026-4d87-a986-b17f02f644d7\" (UID: \"deba5665-b026-4d87-a986-b17f02f644d7\") " Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.339509 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-service-ca" (OuterVolumeSpecName: "service-ca") pod "deba5665-b026-4d87-a986-b17f02f644d7" (UID: "deba5665-b026-4d87-a986-b17f02f644d7"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.341381 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "deba5665-b026-4d87-a986-b17f02f644d7" (UID: "deba5665-b026-4d87-a986-b17f02f644d7"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.341626 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-console-config" (OuterVolumeSpecName: "console-config") pod "deba5665-b026-4d87-a986-b17f02f644d7" (UID: "deba5665-b026-4d87-a986-b17f02f644d7"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.341944 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "deba5665-b026-4d87-a986-b17f02f644d7" (UID: "deba5665-b026-4d87-a986-b17f02f644d7"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.348404 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "deba5665-b026-4d87-a986-b17f02f644d7" (UID: "deba5665-b026-4d87-a986-b17f02f644d7"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.349086 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "deba5665-b026-4d87-a986-b17f02f644d7" (UID: "deba5665-b026-4d87-a986-b17f02f644d7"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.350245 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deba5665-b026-4d87-a986-b17f02f644d7-kube-api-access-ls8gg" (OuterVolumeSpecName: "kube-api-access-ls8gg") pod "deba5665-b026-4d87-a986-b17f02f644d7" (UID: "deba5665-b026-4d87-a986-b17f02f644d7"). InnerVolumeSpecName "kube-api-access-ls8gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.440084 4767 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-console-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.440136 4767 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.440150 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls8gg\" (UniqueName: \"kubernetes.io/projected/deba5665-b026-4d87-a986-b17f02f644d7-kube-api-access-ls8gg\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.440165 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.440194 4767 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.440204 4767 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/deba5665-b026-4d87-a986-b17f02f644d7-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:20 crc kubenswrapper[4767]: I0317 15:45:20.440214 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/deba5665-b026-4d87-a986-b17f02f644d7-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:45:21 crc kubenswrapper[4767]: I0317 15:45:21.064190 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-9cz8q_deba5665-b026-4d87-a986-b17f02f644d7/console/0.log" Mar 17 15:45:21 crc kubenswrapper[4767]: I0317 15:45:21.064270 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9cz8q" event={"ID":"deba5665-b026-4d87-a986-b17f02f644d7","Type":"ContainerDied","Data":"2f730dd27ad571edc2001f46171e6a80713055b782c940d838d8f18ce435558d"} Mar 17 15:45:21 crc kubenswrapper[4767]: I0317 15:45:21.064332 4767 scope.go:117] "RemoveContainer" containerID="415bc454f7b25622d72c606b86b13b8ba7e778cd5cef1a1a64a9fb92f886fbf6" Mar 17 15:45:21 crc kubenswrapper[4767]: I0317 15:45:21.064373 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9cz8q" Mar 17 15:45:21 crc kubenswrapper[4767]: I0317 15:45:21.098108 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-9cz8q"] Mar 17 15:45:21 crc kubenswrapper[4767]: I0317 15:45:21.105661 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-9cz8q"] Mar 17 15:45:21 crc kubenswrapper[4767]: I0317 15:45:21.363903 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deba5665-b026-4d87-a986-b17f02f644d7" path="/var/lib/kubelet/pods/deba5665-b026-4d87-a986-b17f02f644d7/volumes" Mar 17 15:45:24 crc kubenswrapper[4767]: I0317 15:45:24.513293 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:45:24 crc kubenswrapper[4767]: I0317 15:45:24.520900 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" Mar 17 15:45:45 crc kubenswrapper[4767]: I0317 15:45:45.880075 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:45:45 crc kubenswrapper[4767]: I0317 15:45:45.922133 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:45:46 crc kubenswrapper[4767]: I0317 15:45:46.288072 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.542351 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-55c858d47f-r899j"] Mar 17 15:45:55 crc kubenswrapper[4767]: E0317 15:45:55.543769 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83518d71-dc9f-4e6c-9ad4-66ee45c262f8" containerName="registry" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.543794 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="83518d71-dc9f-4e6c-9ad4-66ee45c262f8" containerName="registry" Mar 17 15:45:55 crc kubenswrapper[4767]: E0317 15:45:55.543807 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd451935-a4a7-41df-aa19-8e7065df2b0c" containerName="collect-profiles" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.543814 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd451935-a4a7-41df-aa19-8e7065df2b0c" containerName="collect-profiles" Mar 17 15:45:55 crc kubenswrapper[4767]: E0317 15:45:55.543831 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deba5665-b026-4d87-a986-b17f02f644d7" containerName="console" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.543840 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="deba5665-b026-4d87-a986-b17f02f644d7" containerName="console" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.544014 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd451935-a4a7-41df-aa19-8e7065df2b0c" containerName="collect-profiles" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.544030 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="83518d71-dc9f-4e6c-9ad4-66ee45c262f8" containerName="registry" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.544040 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="deba5665-b026-4d87-a986-b17f02f644d7" containerName="console" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.544792 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.592844 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-55c858d47f-r899j"] Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.666695 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-trusted-ca-bundle\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.666804 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-oauth-config\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.666853 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-config\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.666890 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pkj7\" (UniqueName: \"kubernetes.io/projected/ea4340bd-c8ea-4286-be77-f537b9f0d99d-kube-api-access-2pkj7\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.666922 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-service-ca\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.666967 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-oauth-serving-cert\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.667014 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-serving-cert\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.769700 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-oauth-config\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.769769 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-config\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.769794 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pkj7\" (UniqueName: \"kubernetes.io/projected/ea4340bd-c8ea-4286-be77-f537b9f0d99d-kube-api-access-2pkj7\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.769821 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-service-ca\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.769876 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-oauth-serving-cert\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.769957 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-serving-cert\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.770010 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-trusted-ca-bundle\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.770937 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-config\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.771230 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-service-ca\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.771344 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-oauth-serving-cert\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.771602 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-trusted-ca-bundle\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.777393 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-oauth-config\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.778212 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-serving-cert\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:55 crc kubenswrapper[4767]: I0317 15:45:55.894153 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pkj7\" (UniqueName: \"kubernetes.io/projected/ea4340bd-c8ea-4286-be77-f537b9f0d99d-kube-api-access-2pkj7\") pod \"console-55c858d47f-r899j\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:56 crc kubenswrapper[4767]: I0317 15:45:56.169420 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:45:56 crc kubenswrapper[4767]: I0317 15:45:56.399642 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-55c858d47f-r899j"] Mar 17 15:45:57 crc kubenswrapper[4767]: I0317 15:45:57.342771 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-55c858d47f-r899j" event={"ID":"ea4340bd-c8ea-4286-be77-f537b9f0d99d","Type":"ContainerStarted","Data":"ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b"} Mar 17 15:45:57 crc kubenswrapper[4767]: I0317 15:45:57.342852 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-55c858d47f-r899j" event={"ID":"ea4340bd-c8ea-4286-be77-f537b9f0d99d","Type":"ContainerStarted","Data":"dc6c53fb145841b1b049d8ca462c2a9f40767678638a549527e55092d48df97b"} Mar 17 15:45:57 crc kubenswrapper[4767]: I0317 15:45:57.373797 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-55c858d47f-r899j" podStartSLOduration=2.373766708 podStartE2EDuration="2.373766708s" podCreationTimestamp="2026-03-17 15:45:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:45:57.370517739 +0000 UTC m=+548.783833806" watchObservedRunningTime="2026-03-17 15:45:57.373766708 +0000 UTC m=+548.787082755" Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.148517 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562706-jzpkc"] Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.150540 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562706-jzpkc" Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.153734 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.154016 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.162418 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.168107 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562706-jzpkc"] Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.199096 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqcj7\" (UniqueName: \"kubernetes.io/projected/21009afb-c8b8-4301-a400-8e193d50c6a9-kube-api-access-rqcj7\") pod \"auto-csr-approver-29562706-jzpkc\" (UID: \"21009afb-c8b8-4301-a400-8e193d50c6a9\") " pod="openshift-infra/auto-csr-approver-29562706-jzpkc" Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.299965 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqcj7\" (UniqueName: \"kubernetes.io/projected/21009afb-c8b8-4301-a400-8e193d50c6a9-kube-api-access-rqcj7\") pod \"auto-csr-approver-29562706-jzpkc\" (UID: \"21009afb-c8b8-4301-a400-8e193d50c6a9\") " pod="openshift-infra/auto-csr-approver-29562706-jzpkc" Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.324424 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqcj7\" (UniqueName: \"kubernetes.io/projected/21009afb-c8b8-4301-a400-8e193d50c6a9-kube-api-access-rqcj7\") pod \"auto-csr-approver-29562706-jzpkc\" (UID: \"21009afb-c8b8-4301-a400-8e193d50c6a9\") " pod="openshift-infra/auto-csr-approver-29562706-jzpkc" Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.482560 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562706-jzpkc" Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.851393 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562706-jzpkc"] Mar 17 15:46:00 crc kubenswrapper[4767]: W0317 15:46:00.854728 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21009afb_c8b8_4301_a400_8e193d50c6a9.slice/crio-5fb44f0b3ce3ad13be072f5bb5a24ce139e61aa1e2466ba219af970f3cf82d9d WatchSource:0}: Error finding container 5fb44f0b3ce3ad13be072f5bb5a24ce139e61aa1e2466ba219af970f3cf82d9d: Status 404 returned error can't find the container with id 5fb44f0b3ce3ad13be072f5bb5a24ce139e61aa1e2466ba219af970f3cf82d9d Mar 17 15:46:00 crc kubenswrapper[4767]: I0317 15:46:00.858148 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 15:46:01 crc kubenswrapper[4767]: I0317 15:46:01.369223 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562706-jzpkc" event={"ID":"21009afb-c8b8-4301-a400-8e193d50c6a9","Type":"ContainerStarted","Data":"5fb44f0b3ce3ad13be072f5bb5a24ce139e61aa1e2466ba219af970f3cf82d9d"} Mar 17 15:46:06 crc kubenswrapper[4767]: I0317 15:46:06.169724 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:46:06 crc kubenswrapper[4767]: I0317 15:46:06.170631 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:46:06 crc kubenswrapper[4767]: I0317 15:46:06.176081 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:46:06 crc kubenswrapper[4767]: I0317 15:46:06.408682 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:46:06 crc kubenswrapper[4767]: I0317 15:46:06.576105 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7948d5b58d-tvgzz"] Mar 17 15:46:15 crc kubenswrapper[4767]: I0317 15:46:15.732975 4767 generic.go:334] "Generic (PLEG): container finished" podID="21009afb-c8b8-4301-a400-8e193d50c6a9" containerID="d225fc6960a166ccd176dfa28e583ae0a4c957489c67ff5dc3a46a1a412444c9" exitCode=0 Mar 17 15:46:15 crc kubenswrapper[4767]: I0317 15:46:15.733123 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562706-jzpkc" event={"ID":"21009afb-c8b8-4301-a400-8e193d50c6a9","Type":"ContainerDied","Data":"d225fc6960a166ccd176dfa28e583ae0a4c957489c67ff5dc3a46a1a412444c9"} Mar 17 15:46:17 crc kubenswrapper[4767]: I0317 15:46:17.040049 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562706-jzpkc" Mar 17 15:46:17 crc kubenswrapper[4767]: I0317 15:46:17.190742 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqcj7\" (UniqueName: \"kubernetes.io/projected/21009afb-c8b8-4301-a400-8e193d50c6a9-kube-api-access-rqcj7\") pod \"21009afb-c8b8-4301-a400-8e193d50c6a9\" (UID: \"21009afb-c8b8-4301-a400-8e193d50c6a9\") " Mar 17 15:46:17 crc kubenswrapper[4767]: I0317 15:46:17.200547 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21009afb-c8b8-4301-a400-8e193d50c6a9-kube-api-access-rqcj7" (OuterVolumeSpecName: "kube-api-access-rqcj7") pod "21009afb-c8b8-4301-a400-8e193d50c6a9" (UID: "21009afb-c8b8-4301-a400-8e193d50c6a9"). InnerVolumeSpecName "kube-api-access-rqcj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:46:17 crc kubenswrapper[4767]: I0317 15:46:17.293129 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqcj7\" (UniqueName: \"kubernetes.io/projected/21009afb-c8b8-4301-a400-8e193d50c6a9-kube-api-access-rqcj7\") on node \"crc\" DevicePath \"\"" Mar 17 15:46:17 crc kubenswrapper[4767]: I0317 15:46:17.753539 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562706-jzpkc" event={"ID":"21009afb-c8b8-4301-a400-8e193d50c6a9","Type":"ContainerDied","Data":"5fb44f0b3ce3ad13be072f5bb5a24ce139e61aa1e2466ba219af970f3cf82d9d"} Mar 17 15:46:17 crc kubenswrapper[4767]: I0317 15:46:17.753611 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fb44f0b3ce3ad13be072f5bb5a24ce139e61aa1e2466ba219af970f3cf82d9d" Mar 17 15:46:17 crc kubenswrapper[4767]: I0317 15:46:17.753636 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562706-jzpkc" Mar 17 15:46:18 crc kubenswrapper[4767]: I0317 15:46:18.121045 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562700-h4fl7"] Mar 17 15:46:18 crc kubenswrapper[4767]: I0317 15:46:18.127163 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562700-h4fl7"] Mar 17 15:46:19 crc kubenswrapper[4767]: I0317 15:46:19.365726 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="051f576f-ce90-4401-8944-a340dd67b274" path="/var/lib/kubelet/pods/051f576f-ce90-4401-8944-a340dd67b274/volumes" Mar 17 15:46:31 crc kubenswrapper[4767]: I0317 15:46:31.624655 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-7948d5b58d-tvgzz" podUID="14549ef2-a133-45bf-88b6-6ea5741c6286" containerName="console" containerID="cri-o://7c4221f87ad6df0f0a53c503ef6526d4ab75fb0400000c8ef67df57fc592a371" gracePeriod=15 Mar 17 15:46:31 crc kubenswrapper[4767]: I0317 15:46:31.862778 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7948d5b58d-tvgzz_14549ef2-a133-45bf-88b6-6ea5741c6286/console/0.log" Mar 17 15:46:31 crc kubenswrapper[4767]: I0317 15:46:31.863783 4767 generic.go:334] "Generic (PLEG): container finished" podID="14549ef2-a133-45bf-88b6-6ea5741c6286" containerID="7c4221f87ad6df0f0a53c503ef6526d4ab75fb0400000c8ef67df57fc592a371" exitCode=2 Mar 17 15:46:31 crc kubenswrapper[4767]: I0317 15:46:31.863835 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7948d5b58d-tvgzz" event={"ID":"14549ef2-a133-45bf-88b6-6ea5741c6286","Type":"ContainerDied","Data":"7c4221f87ad6df0f0a53c503ef6526d4ab75fb0400000c8ef67df57fc592a371"} Mar 17 15:46:31 crc kubenswrapper[4767]: I0317 15:46:31.998303 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7948d5b58d-tvgzz_14549ef2-a133-45bf-88b6-6ea5741c6286/console/0.log" Mar 17 15:46:31 crc kubenswrapper[4767]: I0317 15:46:31.998386 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.117876 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzzrq\" (UniqueName: \"kubernetes.io/projected/14549ef2-a133-45bf-88b6-6ea5741c6286-kube-api-access-tzzrq\") pod \"14549ef2-a133-45bf-88b6-6ea5741c6286\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.117984 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-oauth-serving-cert\") pod \"14549ef2-a133-45bf-88b6-6ea5741c6286\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.118035 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-oauth-config\") pod \"14549ef2-a133-45bf-88b6-6ea5741c6286\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.118498 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-console-config\") pod \"14549ef2-a133-45bf-88b6-6ea5741c6286\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.118537 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-trusted-ca-bundle\") pod \"14549ef2-a133-45bf-88b6-6ea5741c6286\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.118614 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-service-ca\") pod \"14549ef2-a133-45bf-88b6-6ea5741c6286\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.118705 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-serving-cert\") pod \"14549ef2-a133-45bf-88b6-6ea5741c6286\" (UID: \"14549ef2-a133-45bf-88b6-6ea5741c6286\") " Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.119108 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-console-config" (OuterVolumeSpecName: "console-config") pod "14549ef2-a133-45bf-88b6-6ea5741c6286" (UID: "14549ef2-a133-45bf-88b6-6ea5741c6286"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.119710 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "14549ef2-a133-45bf-88b6-6ea5741c6286" (UID: "14549ef2-a133-45bf-88b6-6ea5741c6286"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.119738 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "14549ef2-a133-45bf-88b6-6ea5741c6286" (UID: "14549ef2-a133-45bf-88b6-6ea5741c6286"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.119804 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-service-ca" (OuterVolumeSpecName: "service-ca") pod "14549ef2-a133-45bf-88b6-6ea5741c6286" (UID: "14549ef2-a133-45bf-88b6-6ea5741c6286"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.126019 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14549ef2-a133-45bf-88b6-6ea5741c6286-kube-api-access-tzzrq" (OuterVolumeSpecName: "kube-api-access-tzzrq") pod "14549ef2-a133-45bf-88b6-6ea5741c6286" (UID: "14549ef2-a133-45bf-88b6-6ea5741c6286"). InnerVolumeSpecName "kube-api-access-tzzrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.126376 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "14549ef2-a133-45bf-88b6-6ea5741c6286" (UID: "14549ef2-a133-45bf-88b6-6ea5741c6286"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.131410 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "14549ef2-a133-45bf-88b6-6ea5741c6286" (UID: "14549ef2-a133-45bf-88b6-6ea5741c6286"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.220856 4767 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.220906 4767 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.220918 4767 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-console-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.220927 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.220936 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/14549ef2-a133-45bf-88b6-6ea5741c6286-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.220946 4767 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/14549ef2-a133-45bf-88b6-6ea5741c6286-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.220956 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzzrq\" (UniqueName: \"kubernetes.io/projected/14549ef2-a133-45bf-88b6-6ea5741c6286-kube-api-access-tzzrq\") on node \"crc\" DevicePath \"\"" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.874509 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7948d5b58d-tvgzz_14549ef2-a133-45bf-88b6-6ea5741c6286/console/0.log" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.875129 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7948d5b58d-tvgzz" event={"ID":"14549ef2-a133-45bf-88b6-6ea5741c6286","Type":"ContainerDied","Data":"974e083983451ce4d9d44782a7454363b1f66b66f4ae1037b30d038ade9d7e56"} Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.875257 4767 scope.go:117] "RemoveContainer" containerID="7c4221f87ad6df0f0a53c503ef6526d4ab75fb0400000c8ef67df57fc592a371" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.875721 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7948d5b58d-tvgzz" Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.915106 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7948d5b58d-tvgzz"] Mar 17 15:46:32 crc kubenswrapper[4767]: I0317 15:46:32.918800 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-7948d5b58d-tvgzz"] Mar 17 15:46:33 crc kubenswrapper[4767]: I0317 15:46:33.363305 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14549ef2-a133-45bf-88b6-6ea5741c6286" path="/var/lib/kubelet/pods/14549ef2-a133-45bf-88b6-6ea5741c6286/volumes" Mar 17 15:47:30 crc kubenswrapper[4767]: I0317 15:47:30.465434 4767 scope.go:117] "RemoveContainer" containerID="5b246f51d43510d489fa7052349317595bf29b0844bd0914dc6efeb056773172" Mar 17 15:47:34 crc kubenswrapper[4767]: I0317 15:47:34.167398 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:47:34 crc kubenswrapper[4767]: I0317 15:47:34.168457 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.147678 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562708-ghrs9"] Mar 17 15:48:00 crc kubenswrapper[4767]: E0317 15:48:00.148995 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14549ef2-a133-45bf-88b6-6ea5741c6286" containerName="console" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.149014 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="14549ef2-a133-45bf-88b6-6ea5741c6286" containerName="console" Mar 17 15:48:00 crc kubenswrapper[4767]: E0317 15:48:00.149038 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21009afb-c8b8-4301-a400-8e193d50c6a9" containerName="oc" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.149044 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="21009afb-c8b8-4301-a400-8e193d50c6a9" containerName="oc" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.149275 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="21009afb-c8b8-4301-a400-8e193d50c6a9" containerName="oc" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.149315 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="14549ef2-a133-45bf-88b6-6ea5741c6286" containerName="console" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.149890 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562708-ghrs9" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.153848 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.153880 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.153983 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.158264 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562708-ghrs9"] Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.275650 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl7pf\" (UniqueName: \"kubernetes.io/projected/a97b7496-d795-4914-b980-857e83f3f7da-kube-api-access-cl7pf\") pod \"auto-csr-approver-29562708-ghrs9\" (UID: \"a97b7496-d795-4914-b980-857e83f3f7da\") " pod="openshift-infra/auto-csr-approver-29562708-ghrs9" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.378139 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl7pf\" (UniqueName: \"kubernetes.io/projected/a97b7496-d795-4914-b980-857e83f3f7da-kube-api-access-cl7pf\") pod \"auto-csr-approver-29562708-ghrs9\" (UID: \"a97b7496-d795-4914-b980-857e83f3f7da\") " pod="openshift-infra/auto-csr-approver-29562708-ghrs9" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.401291 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl7pf\" (UniqueName: \"kubernetes.io/projected/a97b7496-d795-4914-b980-857e83f3f7da-kube-api-access-cl7pf\") pod \"auto-csr-approver-29562708-ghrs9\" (UID: \"a97b7496-d795-4914-b980-857e83f3f7da\") " pod="openshift-infra/auto-csr-approver-29562708-ghrs9" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.490809 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562708-ghrs9" Mar 17 15:48:00 crc kubenswrapper[4767]: I0317 15:48:00.766441 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562708-ghrs9"] Mar 17 15:48:01 crc kubenswrapper[4767]: I0317 15:48:01.566195 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562708-ghrs9" event={"ID":"a97b7496-d795-4914-b980-857e83f3f7da","Type":"ContainerStarted","Data":"24be26b78de3fc3e1e4a18f082306082b29a6319c26715f5e478850553aaf178"} Mar 17 15:48:02 crc kubenswrapper[4767]: I0317 15:48:02.589924 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562708-ghrs9" event={"ID":"a97b7496-d795-4914-b980-857e83f3f7da","Type":"ContainerStarted","Data":"be17fb93e42627d48a838082c1c40150df0a32372ee56d6c5f699c7fad43a228"} Mar 17 15:48:02 crc kubenswrapper[4767]: I0317 15:48:02.615414 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562708-ghrs9" podStartSLOduration=1.360966301 podStartE2EDuration="2.615385293s" podCreationTimestamp="2026-03-17 15:48:00 +0000 UTC" firstStartedPulling="2026-03-17 15:48:00.779478778 +0000 UTC m=+672.192794825" lastFinishedPulling="2026-03-17 15:48:02.03389778 +0000 UTC m=+673.447213817" observedRunningTime="2026-03-17 15:48:02.612343228 +0000 UTC m=+674.025659285" watchObservedRunningTime="2026-03-17 15:48:02.615385293 +0000 UTC m=+674.028701340" Mar 17 15:48:03 crc kubenswrapper[4767]: I0317 15:48:03.601515 4767 generic.go:334] "Generic (PLEG): container finished" podID="a97b7496-d795-4914-b980-857e83f3f7da" containerID="be17fb93e42627d48a838082c1c40150df0a32372ee56d6c5f699c7fad43a228" exitCode=0 Mar 17 15:48:03 crc kubenswrapper[4767]: I0317 15:48:03.601646 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562708-ghrs9" event={"ID":"a97b7496-d795-4914-b980-857e83f3f7da","Type":"ContainerDied","Data":"be17fb93e42627d48a838082c1c40150df0a32372ee56d6c5f699c7fad43a228"} Mar 17 15:48:04 crc kubenswrapper[4767]: I0317 15:48:04.167156 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:48:04 crc kubenswrapper[4767]: I0317 15:48:04.167827 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:48:04 crc kubenswrapper[4767]: I0317 15:48:04.861880 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562708-ghrs9" Mar 17 15:48:04 crc kubenswrapper[4767]: I0317 15:48:04.960537 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl7pf\" (UniqueName: \"kubernetes.io/projected/a97b7496-d795-4914-b980-857e83f3f7da-kube-api-access-cl7pf\") pod \"a97b7496-d795-4914-b980-857e83f3f7da\" (UID: \"a97b7496-d795-4914-b980-857e83f3f7da\") " Mar 17 15:48:04 crc kubenswrapper[4767]: I0317 15:48:04.969162 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a97b7496-d795-4914-b980-857e83f3f7da-kube-api-access-cl7pf" (OuterVolumeSpecName: "kube-api-access-cl7pf") pod "a97b7496-d795-4914-b980-857e83f3f7da" (UID: "a97b7496-d795-4914-b980-857e83f3f7da"). InnerVolumeSpecName "kube-api-access-cl7pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:48:05 crc kubenswrapper[4767]: I0317 15:48:05.063061 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl7pf\" (UniqueName: \"kubernetes.io/projected/a97b7496-d795-4914-b980-857e83f3f7da-kube-api-access-cl7pf\") on node \"crc\" DevicePath \"\"" Mar 17 15:48:05 crc kubenswrapper[4767]: I0317 15:48:05.620878 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562708-ghrs9" event={"ID":"a97b7496-d795-4914-b980-857e83f3f7da","Type":"ContainerDied","Data":"24be26b78de3fc3e1e4a18f082306082b29a6319c26715f5e478850553aaf178"} Mar 17 15:48:05 crc kubenswrapper[4767]: I0317 15:48:05.620943 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24be26b78de3fc3e1e4a18f082306082b29a6319c26715f5e478850553aaf178" Mar 17 15:48:05 crc kubenswrapper[4767]: I0317 15:48:05.621028 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562708-ghrs9" Mar 17 15:48:05 crc kubenswrapper[4767]: I0317 15:48:05.683111 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562702-hxck5"] Mar 17 15:48:05 crc kubenswrapper[4767]: I0317 15:48:05.689859 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562702-hxck5"] Mar 17 15:48:07 crc kubenswrapper[4767]: I0317 15:48:07.368807 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6670565d-1552-47bf-83c5-bf2cadda7f03" path="/var/lib/kubelet/pods/6670565d-1552-47bf-83c5-bf2cadda7f03/volumes" Mar 17 15:48:30 crc kubenswrapper[4767]: I0317 15:48:30.568010 4767 scope.go:117] "RemoveContainer" containerID="dd01e68ebb275da7b94cc346d8698cbae46ab1a1d65664f44c2feea325983ccf" Mar 17 15:48:34 crc kubenswrapper[4767]: I0317 15:48:34.166969 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:48:34 crc kubenswrapper[4767]: I0317 15:48:34.168074 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:48:34 crc kubenswrapper[4767]: I0317 15:48:34.168237 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:48:34 crc kubenswrapper[4767]: I0317 15:48:34.169406 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d152bc4b2f69796edf7aec43d7f21284777821024a613d49605f7c9d9db18719"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 15:48:34 crc kubenswrapper[4767]: I0317 15:48:34.169494 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://d152bc4b2f69796edf7aec43d7f21284777821024a613d49605f7c9d9db18719" gracePeriod=600 Mar 17 15:48:34 crc kubenswrapper[4767]: I0317 15:48:34.849486 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="d152bc4b2f69796edf7aec43d7f21284777821024a613d49605f7c9d9db18719" exitCode=0 Mar 17 15:48:34 crc kubenswrapper[4767]: I0317 15:48:34.849577 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"d152bc4b2f69796edf7aec43d7f21284777821024a613d49605f7c9d9db18719"} Mar 17 15:48:34 crc kubenswrapper[4767]: I0317 15:48:34.849891 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"942ec751e9132c97b85320d70dbc413daf5abb6e5090f70bbebaf738b7e2ef34"} Mar 17 15:48:34 crc kubenswrapper[4767]: I0317 15:48:34.849924 4767 scope.go:117] "RemoveContainer" containerID="3c904b546ac42362d6e52f6a82d664f6ec63ec5a64d0aa1a9ae36b3788164828" Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.152690 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562710-2bppx"] Mar 17 15:50:00 crc kubenswrapper[4767]: E0317 15:50:00.154226 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a97b7496-d795-4914-b980-857e83f3f7da" containerName="oc" Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.154259 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a97b7496-d795-4914-b980-857e83f3f7da" containerName="oc" Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.154435 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a97b7496-d795-4914-b980-857e83f3f7da" containerName="oc" Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.155205 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562710-2bppx" Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.158718 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.158729 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.158862 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562710-2bppx"] Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.160700 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.215151 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djr87\" (UniqueName: \"kubernetes.io/projected/2f907dc9-1b30-4ae0-8b07-0566948f13bf-kube-api-access-djr87\") pod \"auto-csr-approver-29562710-2bppx\" (UID: \"2f907dc9-1b30-4ae0-8b07-0566948f13bf\") " pod="openshift-infra/auto-csr-approver-29562710-2bppx" Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.316822 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djr87\" (UniqueName: \"kubernetes.io/projected/2f907dc9-1b30-4ae0-8b07-0566948f13bf-kube-api-access-djr87\") pod \"auto-csr-approver-29562710-2bppx\" (UID: \"2f907dc9-1b30-4ae0-8b07-0566948f13bf\") " pod="openshift-infra/auto-csr-approver-29562710-2bppx" Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.344003 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djr87\" (UniqueName: \"kubernetes.io/projected/2f907dc9-1b30-4ae0-8b07-0566948f13bf-kube-api-access-djr87\") pod \"auto-csr-approver-29562710-2bppx\" (UID: \"2f907dc9-1b30-4ae0-8b07-0566948f13bf\") " pod="openshift-infra/auto-csr-approver-29562710-2bppx" Mar 17 15:50:00 crc kubenswrapper[4767]: I0317 15:50:00.486706 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562710-2bppx" Mar 17 15:50:01 crc kubenswrapper[4767]: I0317 15:50:01.245310 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562710-2bppx"] Mar 17 15:50:02 crc kubenswrapper[4767]: I0317 15:50:02.019074 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562710-2bppx" event={"ID":"2f907dc9-1b30-4ae0-8b07-0566948f13bf","Type":"ContainerStarted","Data":"f3d645d3e2889e6ef203fc0361ed7e499110feb9b2950815432b0aafa78e01f2"} Mar 17 15:50:03 crc kubenswrapper[4767]: I0317 15:50:03.089511 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562710-2bppx" event={"ID":"2f907dc9-1b30-4ae0-8b07-0566948f13bf","Type":"ContainerStarted","Data":"32bbcb6ddd2c3090708f57b91d7c71398f21718cabb4c5c0710bdedff7bc91a2"} Mar 17 15:50:03 crc kubenswrapper[4767]: I0317 15:50:03.110258 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562710-2bppx" podStartSLOduration=1.815702363 podStartE2EDuration="3.110232762s" podCreationTimestamp="2026-03-17 15:50:00 +0000 UTC" firstStartedPulling="2026-03-17 15:50:01.261325294 +0000 UTC m=+792.674641341" lastFinishedPulling="2026-03-17 15:50:02.555855693 +0000 UTC m=+793.969171740" observedRunningTime="2026-03-17 15:50:03.107750999 +0000 UTC m=+794.521067046" watchObservedRunningTime="2026-03-17 15:50:03.110232762 +0000 UTC m=+794.523548809" Mar 17 15:50:03 crc kubenswrapper[4767]: I0317 15:50:03.828875 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6"] Mar 17 15:50:03 crc kubenswrapper[4767]: I0317 15:50:03.830589 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:03 crc kubenswrapper[4767]: I0317 15:50:03.833507 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 17 15:50:03 crc kubenswrapper[4767]: I0317 15:50:03.842793 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6"] Mar 17 15:50:04 crc kubenswrapper[4767]: I0317 15:50:04.029941 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:04 crc kubenswrapper[4767]: I0317 15:50:04.029998 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5skrc\" (UniqueName: \"kubernetes.io/projected/ad3da6ba-be8d-4db0-979e-45651941ccd0-kube-api-access-5skrc\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:04 crc kubenswrapper[4767]: I0317 15:50:04.030035 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:04 crc kubenswrapper[4767]: I0317 15:50:04.100837 4767 generic.go:334] "Generic (PLEG): container finished" podID="2f907dc9-1b30-4ae0-8b07-0566948f13bf" containerID="32bbcb6ddd2c3090708f57b91d7c71398f21718cabb4c5c0710bdedff7bc91a2" exitCode=0 Mar 17 15:50:04 crc kubenswrapper[4767]: I0317 15:50:04.101962 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562710-2bppx" event={"ID":"2f907dc9-1b30-4ae0-8b07-0566948f13bf","Type":"ContainerDied","Data":"32bbcb6ddd2c3090708f57b91d7c71398f21718cabb4c5c0710bdedff7bc91a2"} Mar 17 15:50:04 crc kubenswrapper[4767]: I0317 15:50:04.131700 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:04 crc kubenswrapper[4767]: I0317 15:50:04.131790 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5skrc\" (UniqueName: \"kubernetes.io/projected/ad3da6ba-be8d-4db0-979e-45651941ccd0-kube-api-access-5skrc\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:04 crc kubenswrapper[4767]: I0317 15:50:04.131834 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:04 crc kubenswrapper[4767]: I0317 15:50:04.132449 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:04 crc kubenswrapper[4767]: I0317 15:50:04.132488 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:04 crc kubenswrapper[4767]: I0317 15:50:04.155935 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5skrc\" (UniqueName: \"kubernetes.io/projected/ad3da6ba-be8d-4db0-979e-45651941ccd0-kube-api-access-5skrc\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:05 crc kubenswrapper[4767]: I0317 15:50:05.052752 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:05 crc kubenswrapper[4767]: I0317 15:50:05.747151 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562710-2bppx" Mar 17 15:50:05 crc kubenswrapper[4767]: I0317 15:50:05.819492 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6"] Mar 17 15:50:05 crc kubenswrapper[4767]: I0317 15:50:05.853214 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djr87\" (UniqueName: \"kubernetes.io/projected/2f907dc9-1b30-4ae0-8b07-0566948f13bf-kube-api-access-djr87\") pod \"2f907dc9-1b30-4ae0-8b07-0566948f13bf\" (UID: \"2f907dc9-1b30-4ae0-8b07-0566948f13bf\") " Mar 17 15:50:05 crc kubenswrapper[4767]: I0317 15:50:05.860914 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f907dc9-1b30-4ae0-8b07-0566948f13bf-kube-api-access-djr87" (OuterVolumeSpecName: "kube-api-access-djr87") pod "2f907dc9-1b30-4ae0-8b07-0566948f13bf" (UID: "2f907dc9-1b30-4ae0-8b07-0566948f13bf"). InnerVolumeSpecName "kube-api-access-djr87". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:50:05 crc kubenswrapper[4767]: I0317 15:50:05.955240 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djr87\" (UniqueName: \"kubernetes.io/projected/2f907dc9-1b30-4ae0-8b07-0566948f13bf-kube-api-access-djr87\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:06 crc kubenswrapper[4767]: I0317 15:50:06.120713 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" event={"ID":"ad3da6ba-be8d-4db0-979e-45651941ccd0","Type":"ContainerStarted","Data":"b13e2e1b7471b807f86936ce08cbc81245c68452802ebfe9aae831879d0e58eb"} Mar 17 15:50:06 crc kubenswrapper[4767]: I0317 15:50:06.120782 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" event={"ID":"ad3da6ba-be8d-4db0-979e-45651941ccd0","Type":"ContainerStarted","Data":"ae46d5ada7cc41aa4812ece0a726cbb638ef84e757ceca9e9cb7bd34ee36a0e1"} Mar 17 15:50:06 crc kubenswrapper[4767]: I0317 15:50:06.124717 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562710-2bppx" Mar 17 15:50:06 crc kubenswrapper[4767]: I0317 15:50:06.124634 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562710-2bppx" event={"ID":"2f907dc9-1b30-4ae0-8b07-0566948f13bf","Type":"ContainerDied","Data":"f3d645d3e2889e6ef203fc0361ed7e499110feb9b2950815432b0aafa78e01f2"} Mar 17 15:50:06 crc kubenswrapper[4767]: I0317 15:50:06.125210 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3d645d3e2889e6ef203fc0361ed7e499110feb9b2950815432b0aafa78e01f2" Mar 17 15:50:06 crc kubenswrapper[4767]: I0317 15:50:06.309161 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562704-kc8jt"] Mar 17 15:50:06 crc kubenswrapper[4767]: I0317 15:50:06.314995 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562704-kc8jt"] Mar 17 15:50:07 crc kubenswrapper[4767]: I0317 15:50:07.133281 4767 generic.go:334] "Generic (PLEG): container finished" podID="ad3da6ba-be8d-4db0-979e-45651941ccd0" containerID="b13e2e1b7471b807f86936ce08cbc81245c68452802ebfe9aae831879d0e58eb" exitCode=0 Mar 17 15:50:07 crc kubenswrapper[4767]: I0317 15:50:07.133346 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" event={"ID":"ad3da6ba-be8d-4db0-979e-45651941ccd0","Type":"ContainerDied","Data":"b13e2e1b7471b807f86936ce08cbc81245c68452802ebfe9aae831879d0e58eb"} Mar 17 15:50:07 crc kubenswrapper[4767]: I0317 15:50:07.363189 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6747c3a7-dc68-4728-bbe4-f642dae2d4b0" path="/var/lib/kubelet/pods/6747c3a7-dc68-4728-bbe4-f642dae2d4b0/volumes" Mar 17 15:50:09 crc kubenswrapper[4767]: I0317 15:50:09.152099 4767 generic.go:334] "Generic (PLEG): container finished" podID="ad3da6ba-be8d-4db0-979e-45651941ccd0" containerID="fd21b43a87316db79e6dfee54b98d44a764125bfd209afbbd385993def7aa098" exitCode=0 Mar 17 15:50:09 crc kubenswrapper[4767]: I0317 15:50:09.152258 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" event={"ID":"ad3da6ba-be8d-4db0-979e-45651941ccd0","Type":"ContainerDied","Data":"fd21b43a87316db79e6dfee54b98d44a764125bfd209afbbd385993def7aa098"} Mar 17 15:50:10 crc kubenswrapper[4767]: I0317 15:50:10.164024 4767 generic.go:334] "Generic (PLEG): container finished" podID="ad3da6ba-be8d-4db0-979e-45651941ccd0" containerID="404d9fd0bc11c005c0527e9fdbd7dd0bb97544157c2b8c16f5e926be583dfa81" exitCode=0 Mar 17 15:50:10 crc kubenswrapper[4767]: I0317 15:50:10.165284 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" event={"ID":"ad3da6ba-be8d-4db0-979e-45651941ccd0","Type":"ContainerDied","Data":"404d9fd0bc11c005c0527e9fdbd7dd0bb97544157c2b8c16f5e926be583dfa81"} Mar 17 15:50:11 crc kubenswrapper[4767]: I0317 15:50:11.722979 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:11 crc kubenswrapper[4767]: I0317 15:50:11.835883 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-util\") pod \"ad3da6ba-be8d-4db0-979e-45651941ccd0\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " Mar 17 15:50:11 crc kubenswrapper[4767]: I0317 15:50:11.836039 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-bundle\") pod \"ad3da6ba-be8d-4db0-979e-45651941ccd0\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " Mar 17 15:50:11 crc kubenswrapper[4767]: I0317 15:50:11.836126 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5skrc\" (UniqueName: \"kubernetes.io/projected/ad3da6ba-be8d-4db0-979e-45651941ccd0-kube-api-access-5skrc\") pod \"ad3da6ba-be8d-4db0-979e-45651941ccd0\" (UID: \"ad3da6ba-be8d-4db0-979e-45651941ccd0\") " Mar 17 15:50:11 crc kubenswrapper[4767]: I0317 15:50:11.840908 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-bundle" (OuterVolumeSpecName: "bundle") pod "ad3da6ba-be8d-4db0-979e-45651941ccd0" (UID: "ad3da6ba-be8d-4db0-979e-45651941ccd0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:50:11 crc kubenswrapper[4767]: I0317 15:50:11.849892 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad3da6ba-be8d-4db0-979e-45651941ccd0-kube-api-access-5skrc" (OuterVolumeSpecName: "kube-api-access-5skrc") pod "ad3da6ba-be8d-4db0-979e-45651941ccd0" (UID: "ad3da6ba-be8d-4db0-979e-45651941ccd0"). InnerVolumeSpecName "kube-api-access-5skrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:50:11 crc kubenswrapper[4767]: I0317 15:50:11.938547 4767 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:11 crc kubenswrapper[4767]: I0317 15:50:11.938606 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5skrc\" (UniqueName: \"kubernetes.io/projected/ad3da6ba-be8d-4db0-979e-45651941ccd0-kube-api-access-5skrc\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:12 crc kubenswrapper[4767]: I0317 15:50:12.014456 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-util" (OuterVolumeSpecName: "util") pod "ad3da6ba-be8d-4db0-979e-45651941ccd0" (UID: "ad3da6ba-be8d-4db0-979e-45651941ccd0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:50:12 crc kubenswrapper[4767]: I0317 15:50:12.040756 4767 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad3da6ba-be8d-4db0-979e-45651941ccd0-util\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:12 crc kubenswrapper[4767]: I0317 15:50:12.185768 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" event={"ID":"ad3da6ba-be8d-4db0-979e-45651941ccd0","Type":"ContainerDied","Data":"ae46d5ada7cc41aa4812ece0a726cbb638ef84e757ceca9e9cb7bd34ee36a0e1"} Mar 17 15:50:12 crc kubenswrapper[4767]: I0317 15:50:12.185865 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae46d5ada7cc41aa4812ece0a726cbb638ef84e757ceca9e9cb7bd34ee36a0e1" Mar 17 15:50:12 crc kubenswrapper[4767]: I0317 15:50:12.185910 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf92397266sqs6" Mar 17 15:50:13 crc kubenswrapper[4767]: I0317 15:50:13.157337 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c4n7r"] Mar 17 15:50:13 crc kubenswrapper[4767]: I0317 15:50:13.158008 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovn-controller" containerID="cri-o://f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133" gracePeriod=30 Mar 17 15:50:13 crc kubenswrapper[4767]: I0317 15:50:13.158101 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="nbdb" containerID="cri-o://b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7" gracePeriod=30 Mar 17 15:50:13 crc kubenswrapper[4767]: I0317 15:50:13.158207 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="northd" containerID="cri-o://28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447" gracePeriod=30 Mar 17 15:50:13 crc kubenswrapper[4767]: I0317 15:50:13.158203 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde" gracePeriod=30 Mar 17 15:50:13 crc kubenswrapper[4767]: I0317 15:50:13.158250 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="kube-rbac-proxy-node" containerID="cri-o://8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a" gracePeriod=30 Mar 17 15:50:13 crc kubenswrapper[4767]: I0317 15:50:13.158309 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="sbdb" containerID="cri-o://b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1" gracePeriod=30 Mar 17 15:50:13 crc kubenswrapper[4767]: I0317 15:50:13.158277 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovn-acl-logging" containerID="cri-o://f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615" gracePeriod=30 Mar 17 15:50:13 crc kubenswrapper[4767]: I0317 15:50:13.270294 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" containerID="cri-o://01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f" gracePeriod=30 Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.257092 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-v6gnj_f36218dd-ddd3-4269-af32-7fe03113b342/kube-multus/1.log" Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.258011 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-v6gnj_f36218dd-ddd3-4269-af32-7fe03113b342/kube-multus/0.log" Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.258052 4767 generic.go:334] "Generic (PLEG): container finished" podID="f36218dd-ddd3-4269-af32-7fe03113b342" containerID="450ee51b49ffbd25e250457c010af40e72c339e723734bcd701ed853117b9041" exitCode=2 Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.258118 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-v6gnj" event={"ID":"f36218dd-ddd3-4269-af32-7fe03113b342","Type":"ContainerDied","Data":"450ee51b49ffbd25e250457c010af40e72c339e723734bcd701ed853117b9041"} Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.258164 4767 scope.go:117] "RemoveContainer" containerID="6bfe053b143473b4194a65db46065fb126cf4a8e98a14e8d8bbfca9c1fa0b47e" Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.259530 4767 scope.go:117] "RemoveContainer" containerID="450ee51b49ffbd25e250457c010af40e72c339e723734bcd701ed853117b9041" Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.260979 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovnkube-controller/2.log" Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.264773 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovn-acl-logging/0.log" Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.265350 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovn-controller/0.log" Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266413 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f" exitCode=0 Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266438 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1" exitCode=0 Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266447 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7" exitCode=0 Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266456 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447" exitCode=0 Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266464 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615" exitCode=143 Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266473 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133" exitCode=143 Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266496 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f"} Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266527 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1"} Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266539 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7"} Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266550 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447"} Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266561 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615"} Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.266570 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133"} Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.314635 4767 scope.go:117] "RemoveContainer" containerID="544bd012a7f9b44c0412a49fdcf29e2de36600e44fe5e236740f022e4332624b" Mar 17 15:50:14 crc kubenswrapper[4767]: E0317 15:50:14.531473 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7 is running failed: container process not found" containerID="b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Mar 17 15:50:14 crc kubenswrapper[4767]: E0317 15:50:14.531610 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1 is running failed: container process not found" containerID="b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Mar 17 15:50:14 crc kubenswrapper[4767]: E0317 15:50:14.532127 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7 is running failed: container process not found" containerID="b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Mar 17 15:50:14 crc kubenswrapper[4767]: E0317 15:50:14.532154 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1 is running failed: container process not found" containerID="b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Mar 17 15:50:14 crc kubenswrapper[4767]: E0317 15:50:14.532478 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7 is running failed: container process not found" containerID="b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Mar 17 15:50:14 crc kubenswrapper[4767]: E0317 15:50:14.532678 4767 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="nbdb" Mar 17 15:50:14 crc kubenswrapper[4767]: E0317 15:50:14.532634 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1 is running failed: container process not found" containerID="b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Mar 17 15:50:14 crc kubenswrapper[4767]: E0317 15:50:14.532956 4767 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="sbdb" Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.984231 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovn-acl-logging/0.log" Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.985600 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovn-controller/0.log" Mar 17 15:50:14 crc kubenswrapper[4767]: I0317 15:50:14.986190 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075118 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rh9xl"] Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075432 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="kube-rbac-proxy-node" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075447 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="kube-rbac-proxy-node" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075458 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="kube-rbac-proxy-ovn-metrics" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075464 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="kube-rbac-proxy-ovn-metrics" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075472 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075479 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075489 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad3da6ba-be8d-4db0-979e-45651941ccd0" containerName="extract" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075497 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad3da6ba-be8d-4db0-979e-45651941ccd0" containerName="extract" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075506 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="kubecfg-setup" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075512 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="kubecfg-setup" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075522 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="northd" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075527 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="northd" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075538 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="sbdb" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075543 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="sbdb" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075550 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad3da6ba-be8d-4db0-979e-45651941ccd0" containerName="pull" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075555 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad3da6ba-be8d-4db0-979e-45651941ccd0" containerName="pull" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075565 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="nbdb" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075570 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="nbdb" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075580 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075585 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075595 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad3da6ba-be8d-4db0-979e-45651941ccd0" containerName="util" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075601 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad3da6ba-be8d-4db0-979e-45651941ccd0" containerName="util" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075615 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovn-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075621 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovn-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075629 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f907dc9-1b30-4ae0-8b07-0566948f13bf" containerName="oc" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075635 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f907dc9-1b30-4ae0-8b07-0566948f13bf" containerName="oc" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075645 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovn-acl-logging" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075650 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovn-acl-logging" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075752 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075761 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075767 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075777 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="northd" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075785 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="sbdb" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075795 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovn-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075804 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="kube-rbac-proxy-node" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075813 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f907dc9-1b30-4ae0-8b07-0566948f13bf" containerName="oc" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075818 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad3da6ba-be8d-4db0-979e-45651941ccd0" containerName="extract" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075826 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="kube-rbac-proxy-ovn-metrics" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075835 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovn-acl-logging" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075842 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="nbdb" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.075931 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.075939 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.076045 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.076142 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.076149 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerName="ovnkube-controller" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.078060 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.089500 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpdb2\" (UniqueName: \"kubernetes.io/projected/93753cb4-c3fa-424e-84a2-0a7db20eecfa-kube-api-access-zpdb2\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090419 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-netns\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090480 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-slash\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090522 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-etc-openvswitch\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090582 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090618 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-env-overrides\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090622 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-slash" (OuterVolumeSpecName: "host-slash") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090654 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090655 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-script-lib\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090684 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-bin\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090754 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-systemd\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090791 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-openvswitch\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090822 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovn-node-metrics-cert\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090839 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090859 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-ovn\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090952 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091025 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091003 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.090994 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-netd\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091225 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-var-lib-openvswitch\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091262 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-ovn-kubernetes\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091269 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091300 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-kubelet\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091308 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091334 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091378 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-config\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091417 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-var-lib-cni-networks-ovn-kubernetes\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091467 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-log-socket\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091488 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-systemd-units\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091511 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-node-log\") pod \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\" (UID: \"93753cb4-c3fa-424e-84a2-0a7db20eecfa\") " Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091550 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091590 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091616 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091684 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-log-socket" (OuterVolumeSpecName: "log-socket") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091730 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091757 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-node-log" (OuterVolumeSpecName: "node-log") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.091885 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092186 4767 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092203 4767 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092216 4767 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-log-socket\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092225 4767 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092235 4767 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-node-log\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092244 4767 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092253 4767 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-slash\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092261 4767 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092271 4767 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092279 4767 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092288 4767 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092297 4767 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092306 4767 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092313 4767 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092321 4767 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092332 4767 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.092340 4767 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.097682 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.122072 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93753cb4-c3fa-424e-84a2-0a7db20eecfa-kube-api-access-zpdb2" (OuterVolumeSpecName: "kube-api-access-zpdb2") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "kube-api-access-zpdb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.141771 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "93753cb4-c3fa-424e-84a2-0a7db20eecfa" (UID: "93753cb4-c3fa-424e-84a2-0a7db20eecfa"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.193795 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-var-lib-openvswitch\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.193875 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/259ec130-5ec4-4841-b227-04b1b2b4134d-ovnkube-script-lib\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.193907 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-cni-bin\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.193938 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-run-ovn-kubernetes\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.193965 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/259ec130-5ec4-4841-b227-04b1b2b4134d-ovn-node-metrics-cert\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194005 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-run-ovn\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194036 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-node-log\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194076 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-run-openvswitch\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194118 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-run-systemd\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194144 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-log-socket\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194196 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/259ec130-5ec4-4841-b227-04b1b2b4134d-env-overrides\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194223 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw6cq\" (UniqueName: \"kubernetes.io/projected/259ec130-5ec4-4841-b227-04b1b2b4134d-kube-api-access-jw6cq\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194255 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-etc-openvswitch\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194291 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194335 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-slash\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194373 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-systemd-units\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194432 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/259ec130-5ec4-4841-b227-04b1b2b4134d-ovnkube-config\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194489 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-run-netns\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194524 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-kubelet\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194551 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-cni-netd\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194605 4767 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93753cb4-c3fa-424e-84a2-0a7db20eecfa-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194621 4767 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93753cb4-c3fa-424e-84a2-0a7db20eecfa-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.194643 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpdb2\" (UniqueName: \"kubernetes.io/projected/93753cb4-c3fa-424e-84a2-0a7db20eecfa-kube-api-access-zpdb2\") on node \"crc\" DevicePath \"\"" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.280234 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-v6gnj_f36218dd-ddd3-4269-af32-7fe03113b342/kube-multus/1.log" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.280372 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-v6gnj" event={"ID":"f36218dd-ddd3-4269-af32-7fe03113b342","Type":"ContainerStarted","Data":"8f09f98ac71bb81a4bb7dfb1c80500cd0809b56a138316c9570d23ebc3749dba"} Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.286817 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovn-acl-logging/0.log" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.287626 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c4n7r_93753cb4-c3fa-424e-84a2-0a7db20eecfa/ovn-controller/0.log" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.288020 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde" exitCode=0 Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.288044 4767 generic.go:334] "Generic (PLEG): container finished" podID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" containerID="8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a" exitCode=0 Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.288085 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde"} Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.288208 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a"} Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.288229 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" event={"ID":"93753cb4-c3fa-424e-84a2-0a7db20eecfa","Type":"ContainerDied","Data":"9d31966e2e47c089ff7da20e99a504f1676eac18a67ebc2a715662046186f2c6"} Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.288285 4767 scope.go:117] "RemoveContainer" containerID="01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.288717 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c4n7r" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296333 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/259ec130-5ec4-4841-b227-04b1b2b4134d-env-overrides\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296433 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-etc-openvswitch\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296466 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw6cq\" (UniqueName: \"kubernetes.io/projected/259ec130-5ec4-4841-b227-04b1b2b4134d-kube-api-access-jw6cq\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296492 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296536 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-slash\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296586 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-systemd-units\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296622 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/259ec130-5ec4-4841-b227-04b1b2b4134d-ovnkube-config\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296655 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-run-netns\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296695 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-kubelet\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296722 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-cni-netd\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296756 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-var-lib-openvswitch\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296781 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/259ec130-5ec4-4841-b227-04b1b2b4134d-ovnkube-script-lib\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296805 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-cni-bin\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296838 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-run-ovn-kubernetes\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296865 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/259ec130-5ec4-4841-b227-04b1b2b4134d-ovn-node-metrics-cert\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296892 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-run-ovn\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296928 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-node-log\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.296970 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-run-openvswitch\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.297019 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-run-systemd\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.297055 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-log-socket\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.297157 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-log-socket\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.297162 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/259ec130-5ec4-4841-b227-04b1b2b4134d-env-overrides\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.297232 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-etc-openvswitch\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.297277 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-var-lib-openvswitch\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.297726 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.297773 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-slash\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.297860 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-systemd-units\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.298078 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/259ec130-5ec4-4841-b227-04b1b2b4134d-ovnkube-script-lib\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.298147 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-cni-bin\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.298211 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-run-ovn-kubernetes\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.298775 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/259ec130-5ec4-4841-b227-04b1b2b4134d-ovnkube-config\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.298834 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-run-netns\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.298872 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-kubelet\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.298912 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-host-cni-netd\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.298949 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-node-log\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.298981 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-run-ovn\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.299011 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-run-openvswitch\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.299049 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/259ec130-5ec4-4841-b227-04b1b2b4134d-run-systemd\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.304825 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/259ec130-5ec4-4841-b227-04b1b2b4134d-ovn-node-metrics-cert\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.322408 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw6cq\" (UniqueName: \"kubernetes.io/projected/259ec130-5ec4-4841-b227-04b1b2b4134d-kube-api-access-jw6cq\") pod \"ovnkube-node-rh9xl\" (UID: \"259ec130-5ec4-4841-b227-04b1b2b4134d\") " pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.335930 4767 scope.go:117] "RemoveContainer" containerID="b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.349419 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c4n7r"] Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.353002 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c4n7r"] Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.365139 4767 scope.go:117] "RemoveContainer" containerID="b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.365370 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93753cb4-c3fa-424e-84a2-0a7db20eecfa" path="/var/lib/kubelet/pods/93753cb4-c3fa-424e-84a2-0a7db20eecfa/volumes" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.399267 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.399630 4767 scope.go:117] "RemoveContainer" containerID="28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.421302 4767 scope.go:117] "RemoveContainer" containerID="fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde" Mar 17 15:50:15 crc kubenswrapper[4767]: W0317 15:50:15.433259 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod259ec130_5ec4_4841_b227_04b1b2b4134d.slice/crio-0ed4978617233bf62ff79c8ba88bc78eb28c82fe212ef297e44e7aedb5df5477 WatchSource:0}: Error finding container 0ed4978617233bf62ff79c8ba88bc78eb28c82fe212ef297e44e7aedb5df5477: Status 404 returned error can't find the container with id 0ed4978617233bf62ff79c8ba88bc78eb28c82fe212ef297e44e7aedb5df5477 Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.481326 4767 scope.go:117] "RemoveContainer" containerID="8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.516831 4767 scope.go:117] "RemoveContainer" containerID="f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.538466 4767 scope.go:117] "RemoveContainer" containerID="f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.561879 4767 scope.go:117] "RemoveContainer" containerID="223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.585426 4767 scope.go:117] "RemoveContainer" containerID="01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.586128 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f\": container with ID starting with 01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f not found: ID does not exist" containerID="01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.586203 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f"} err="failed to get container status \"01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f\": rpc error: code = NotFound desc = could not find container \"01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f\": container with ID starting with 01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.586243 4767 scope.go:117] "RemoveContainer" containerID="b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.586770 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\": container with ID starting with b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1 not found: ID does not exist" containerID="b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.586789 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1"} err="failed to get container status \"b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\": rpc error: code = NotFound desc = could not find container \"b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\": container with ID starting with b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1 not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.586813 4767 scope.go:117] "RemoveContainer" containerID="b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.587089 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\": container with ID starting with b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7 not found: ID does not exist" containerID="b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.587120 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7"} err="failed to get container status \"b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\": rpc error: code = NotFound desc = could not find container \"b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\": container with ID starting with b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7 not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.587136 4767 scope.go:117] "RemoveContainer" containerID="28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.587551 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\": container with ID starting with 28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447 not found: ID does not exist" containerID="28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.587576 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447"} err="failed to get container status \"28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\": rpc error: code = NotFound desc = could not find container \"28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\": container with ID starting with 28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447 not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.587590 4767 scope.go:117] "RemoveContainer" containerID="fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.587883 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\": container with ID starting with fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde not found: ID does not exist" containerID="fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.587940 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde"} err="failed to get container status \"fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\": rpc error: code = NotFound desc = could not find container \"fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\": container with ID starting with fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.587981 4767 scope.go:117] "RemoveContainer" containerID="8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.588274 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\": container with ID starting with 8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a not found: ID does not exist" containerID="8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.588305 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a"} err="failed to get container status \"8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\": rpc error: code = NotFound desc = could not find container \"8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\": container with ID starting with 8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.588337 4767 scope.go:117] "RemoveContainer" containerID="f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.588675 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\": container with ID starting with f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615 not found: ID does not exist" containerID="f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.588727 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615"} err="failed to get container status \"f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\": rpc error: code = NotFound desc = could not find container \"f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\": container with ID starting with f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615 not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.588742 4767 scope.go:117] "RemoveContainer" containerID="f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.589060 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\": container with ID starting with f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133 not found: ID does not exist" containerID="f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.589097 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133"} err="failed to get container status \"f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\": rpc error: code = NotFound desc = could not find container \"f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\": container with ID starting with f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133 not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.589117 4767 scope.go:117] "RemoveContainer" containerID="223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff" Mar 17 15:50:15 crc kubenswrapper[4767]: E0317 15:50:15.589543 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\": container with ID starting with 223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff not found: ID does not exist" containerID="223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.589579 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff"} err="failed to get container status \"223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\": rpc error: code = NotFound desc = could not find container \"223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\": container with ID starting with 223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.589600 4767 scope.go:117] "RemoveContainer" containerID="01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.589942 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f"} err="failed to get container status \"01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f\": rpc error: code = NotFound desc = could not find container \"01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f\": container with ID starting with 01ffcecde75d55b97aa38e5f78f73235655ba8bad72574585a5b78f1c8d5062f not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.589971 4767 scope.go:117] "RemoveContainer" containerID="b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.590432 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1"} err="failed to get container status \"b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\": rpc error: code = NotFound desc = could not find container \"b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1\": container with ID starting with b1b3e5c86fcabe5ac71c55f2752f6e68423ef10996d157ad8f3440f5e5f8f7d1 not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.590474 4767 scope.go:117] "RemoveContainer" containerID="b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.590757 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7"} err="failed to get container status \"b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\": rpc error: code = NotFound desc = could not find container \"b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7\": container with ID starting with b1f032f16f082675c166e6ef246aa56a7f7b551fc43e40d708452c602f9d98f7 not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.590780 4767 scope.go:117] "RemoveContainer" containerID="28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.591107 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447"} err="failed to get container status \"28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\": rpc error: code = NotFound desc = could not find container \"28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447\": container with ID starting with 28a989c90c01efbdef882e61e14cb26025c46aebb87df5585e7dd1977c5e3447 not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.591165 4767 scope.go:117] "RemoveContainer" containerID="fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.591602 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde"} err="failed to get container status \"fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\": rpc error: code = NotFound desc = could not find container \"fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde\": container with ID starting with fd5c9cb44e61c99062277550b01f515fced6f1aba950cbee990afee4bb7fefde not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.591654 4767 scope.go:117] "RemoveContainer" containerID="8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.591855 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a"} err="failed to get container status \"8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\": rpc error: code = NotFound desc = could not find container \"8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a\": container with ID starting with 8af5859be1a97d90ed0c147526be5411888bd540ec0a4a5d090dcf610ed9fe9a not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.592179 4767 scope.go:117] "RemoveContainer" containerID="f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.592576 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615"} err="failed to get container status \"f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\": rpc error: code = NotFound desc = could not find container \"f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615\": container with ID starting with f053eb2d2156923181e3ecbf8924ae451fa9545aa81826d6501ad3d79453c615 not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.592606 4767 scope.go:117] "RemoveContainer" containerID="f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.592862 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133"} err="failed to get container status \"f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\": rpc error: code = NotFound desc = could not find container \"f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133\": container with ID starting with f8cfb0ec1d7926d00a8e8efd0abb0a45b9250024b1bfcc384b024a0cba199133 not found: ID does not exist" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.592893 4767 scope.go:117] "RemoveContainer" containerID="223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff" Mar 17 15:50:15 crc kubenswrapper[4767]: I0317 15:50:15.593152 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff"} err="failed to get container status \"223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\": rpc error: code = NotFound desc = could not find container \"223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff\": container with ID starting with 223c52de3cdbe79e1310587a3eeaa17a4b60be6c66f69ebcc4a6f6342ff194ff not found: ID does not exist" Mar 17 15:50:16 crc kubenswrapper[4767]: I0317 15:50:16.296290 4767 generic.go:334] "Generic (PLEG): container finished" podID="259ec130-5ec4-4841-b227-04b1b2b4134d" containerID="d134334bc58ed9ce4f6b213f37339ada547193f12f9ec0fa92b162e396f7fd46" exitCode=0 Mar 17 15:50:16 crc kubenswrapper[4767]: I0317 15:50:16.296376 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" event={"ID":"259ec130-5ec4-4841-b227-04b1b2b4134d","Type":"ContainerDied","Data":"d134334bc58ed9ce4f6b213f37339ada547193f12f9ec0fa92b162e396f7fd46"} Mar 17 15:50:16 crc kubenswrapper[4767]: I0317 15:50:16.296420 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" event={"ID":"259ec130-5ec4-4841-b227-04b1b2b4134d","Type":"ContainerStarted","Data":"0ed4978617233bf62ff79c8ba88bc78eb28c82fe212ef297e44e7aedb5df5477"} Mar 17 15:50:17 crc kubenswrapper[4767]: I0317 15:50:17.331331 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" event={"ID":"259ec130-5ec4-4841-b227-04b1b2b4134d","Type":"ContainerStarted","Data":"82af4693278c80e44c1c933a53798e360df33d9a47080bbae0db3fb478a0d984"} Mar 17 15:50:17 crc kubenswrapper[4767]: I0317 15:50:17.333229 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" event={"ID":"259ec130-5ec4-4841-b227-04b1b2b4134d","Type":"ContainerStarted","Data":"c28b141be2a2dee567a3532f5613717475024697e293c0d55a36941bb23e3a59"} Mar 17 15:50:17 crc kubenswrapper[4767]: I0317 15:50:17.333267 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" event={"ID":"259ec130-5ec4-4841-b227-04b1b2b4134d","Type":"ContainerStarted","Data":"bde2e8494e5b887e3a2a8430d8689675d32ff7e1cfaaf6d01ffc933659dcb996"} Mar 17 15:50:17 crc kubenswrapper[4767]: I0317 15:50:17.333285 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" event={"ID":"259ec130-5ec4-4841-b227-04b1b2b4134d","Type":"ContainerStarted","Data":"2227e192f0e1db8645e2ab7e9f7da960f9fb66b40f7a511cd1a9036d27881366"} Mar 17 15:50:18 crc kubenswrapper[4767]: I0317 15:50:18.345524 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" event={"ID":"259ec130-5ec4-4841-b227-04b1b2b4134d","Type":"ContainerStarted","Data":"d1224cb5f6e1cad648c3d2ecac25f31eafcce766ff4f3f331529460bda84b7ed"} Mar 17 15:50:18 crc kubenswrapper[4767]: I0317 15:50:18.345594 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" event={"ID":"259ec130-5ec4-4841-b227-04b1b2b4134d","Type":"ContainerStarted","Data":"3aa5aee7a6f5c7e68e289830fea5b7a3b86b8d8292f270ff0ad931bda1a650b7"} Mar 17 15:50:20 crc kubenswrapper[4767]: I0317 15:50:20.423773 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" event={"ID":"259ec130-5ec4-4841-b227-04b1b2b4134d","Type":"ContainerStarted","Data":"9bfeb39bc6a8686b189c810d313d9591b8f53fef94c47c8b8e9f8405454bd48c"} Mar 17 15:50:23 crc kubenswrapper[4767]: I0317 15:50:23.638018 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" event={"ID":"259ec130-5ec4-4841-b227-04b1b2b4134d","Type":"ContainerStarted","Data":"b11105ad1fede3746da819597a1b2e57bd7d3759c1ace5fd5871f0c5d9678138"} Mar 17 15:50:23 crc kubenswrapper[4767]: I0317 15:50:23.640700 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:23 crc kubenswrapper[4767]: I0317 15:50:23.640731 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:23 crc kubenswrapper[4767]: I0317 15:50:23.640778 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:23 crc kubenswrapper[4767]: I0317 15:50:23.792349 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:24 crc kubenswrapper[4767]: I0317 15:50:24.030292 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:26 crc kubenswrapper[4767]: I0317 15:50:26.540018 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" podStartSLOduration=11.53997526 podStartE2EDuration="11.53997526s" podCreationTimestamp="2026-03-17 15:50:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:50:25.800682905 +0000 UTC m=+817.213998962" watchObservedRunningTime="2026-03-17 15:50:26.53997526 +0000 UTC m=+817.953291307" Mar 17 15:50:27 crc kubenswrapper[4767]: I0317 15:50:27.175227 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" podUID="259ec130-5ec4-4841-b227-04b1b2b4134d" containerName="ovnkube-controller" probeResult="failure" output="" Mar 17 15:50:30 crc kubenswrapper[4767]: I0317 15:50:30.675536 4767 scope.go:117] "RemoveContainer" containerID="22f7c90dd5848ad907327e3b3134bde167e737517b6f81ee131d64d6080bea08" Mar 17 15:50:32 crc kubenswrapper[4767]: I0317 15:50:32.304621 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6"] Mar 17 15:50:32 crc kubenswrapper[4767]: I0317 15:50:32.308856 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:32 crc kubenswrapper[4767]: I0317 15:50:32.315765 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-6nn25" Mar 17 15:50:32 crc kubenswrapper[4767]: I0317 15:50:32.318797 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Mar 17 15:50:32 crc kubenswrapper[4767]: I0317 15:50:32.319435 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Mar 17 15:50:32 crc kubenswrapper[4767]: I0317 15:50:32.577006 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmbfv\" (UniqueName: \"kubernetes.io/projected/1373b043-07be-4e46-9913-4ba92b42352b-kube-api-access-dmbfv\") pod \"obo-prometheus-operator-8ff7d675-j8nb6\" (UID: \"1373b043-07be-4e46-9913-4ba92b42352b\") " pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:32 crc kubenswrapper[4767]: I0317 15:50:32.656003 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6"] Mar 17 15:50:32 crc kubenswrapper[4767]: I0317 15:50:32.680296 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmbfv\" (UniqueName: \"kubernetes.io/projected/1373b043-07be-4e46-9913-4ba92b42352b-kube-api-access-dmbfv\") pod \"obo-prometheus-operator-8ff7d675-j8nb6\" (UID: \"1373b043-07be-4e46-9913-4ba92b42352b\") " pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:32 crc kubenswrapper[4767]: I0317 15:50:32.712826 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmbfv\" (UniqueName: \"kubernetes.io/projected/1373b043-07be-4e46-9913-4ba92b42352b-kube-api-access-dmbfv\") pod \"obo-prometheus-operator-8ff7d675-j8nb6\" (UID: \"1373b043-07be-4e46-9913-4ba92b42352b\") " pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:32 crc kubenswrapper[4767]: I0317 15:50:32.880027 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:32 crc kubenswrapper[4767]: E0317 15:50:32.962069 4767 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators_1373b043-07be-4e46-9913-4ba92b42352b_0(e0659adcbcdb8327b1d3550c50dd14edf857976ab761f012fb77671013ce53c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:50:32 crc kubenswrapper[4767]: E0317 15:50:32.962636 4767 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators_1373b043-07be-4e46-9913-4ba92b42352b_0(e0659adcbcdb8327b1d3550c50dd14edf857976ab761f012fb77671013ce53c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:32 crc kubenswrapper[4767]: E0317 15:50:32.962667 4767 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators_1373b043-07be-4e46-9913-4ba92b42352b_0(e0659adcbcdb8327b1d3550c50dd14edf857976ab761f012fb77671013ce53c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:32 crc kubenswrapper[4767]: E0317 15:50:32.962734 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators(1373b043-07be-4e46-9913-4ba92b42352b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators(1373b043-07be-4e46-9913-4ba92b42352b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators_1373b043-07be-4e46-9913-4ba92b42352b_0(e0659adcbcdb8327b1d3550c50dd14edf857976ab761f012fb77671013ce53c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" podUID="1373b043-07be-4e46-9913-4ba92b42352b" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.418528 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t"] Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.419903 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.427767 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.428341 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-7669b" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.453958 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t"] Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.468095 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp"] Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.469425 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.481149 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp"] Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.568787 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aa728909-9d71-4e9c-a07c-e427969aa381-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t\" (UID: \"aa728909-9d71-4e9c-a07c-e427969aa381\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.569152 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aa728909-9d71-4e9c-a07c-e427969aa381-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t\" (UID: \"aa728909-9d71-4e9c-a07c-e427969aa381\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.597061 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.597665 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:33 crc kubenswrapper[4767]: E0317 15:50:33.636545 4767 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators_1373b043-07be-4e46-9913-4ba92b42352b_0(4fba6d58a02c22994f4a7e07c1fb4c2b776fdecd4e04f3c170abab062d0b56b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 15:50:33 crc kubenswrapper[4767]: E0317 15:50:33.636634 4767 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators_1373b043-07be-4e46-9913-4ba92b42352b_0(4fba6d58a02c22994f4a7e07c1fb4c2b776fdecd4e04f3c170abab062d0b56b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:33 crc kubenswrapper[4767]: E0317 15:50:33.636663 4767 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators_1373b043-07be-4e46-9913-4ba92b42352b_0(4fba6d58a02c22994f4a7e07c1fb4c2b776fdecd4e04f3c170abab062d0b56b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:33 crc kubenswrapper[4767]: E0317 15:50:33.636702 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators(1373b043-07be-4e46-9913-4ba92b42352b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators(1373b043-07be-4e46-9913-4ba92b42352b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-8ff7d675-j8nb6_openshift-operators_1373b043-07be-4e46-9913-4ba92b42352b_0(4fba6d58a02c22994f4a7e07c1fb4c2b776fdecd4e04f3c170abab062d0b56b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" podUID="1373b043-07be-4e46-9913-4ba92b42352b" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.671923 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aa728909-9d71-4e9c-a07c-e427969aa381-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t\" (UID: \"aa728909-9d71-4e9c-a07c-e427969aa381\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.671989 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0ed10882-f7ce-4f57-bb6b-b70b8eac5340-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp\" (UID: \"0ed10882-f7ce-4f57-bb6b-b70b8eac5340\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.672028 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0ed10882-f7ce-4f57-bb6b-b70b8eac5340-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp\" (UID: \"0ed10882-f7ce-4f57-bb6b-b70b8eac5340\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.672118 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aa728909-9d71-4e9c-a07c-e427969aa381-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t\" (UID: \"aa728909-9d71-4e9c-a07c-e427969aa381\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.678876 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aa728909-9d71-4e9c-a07c-e427969aa381-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t\" (UID: \"aa728909-9d71-4e9c-a07c-e427969aa381\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.679036 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aa728909-9d71-4e9c-a07c-e427969aa381-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t\" (UID: \"aa728909-9d71-4e9c-a07c-e427969aa381\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.734693 4767 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.754765 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.766884 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-6dd7dd855f-hw4fk"] Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.775789 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0ed10882-f7ce-4f57-bb6b-b70b8eac5340-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp\" (UID: \"0ed10882-f7ce-4f57-bb6b-b70b8eac5340\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.775844 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0ed10882-f7ce-4f57-bb6b-b70b8eac5340-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp\" (UID: \"0ed10882-f7ce-4f57-bb6b-b70b8eac5340\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.776497 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.785049 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.785406 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-bzmzq" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.786352 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0ed10882-f7ce-4f57-bb6b-b70b8eac5340-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp\" (UID: \"0ed10882-f7ce-4f57-bb6b-b70b8eac5340\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.793030 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0ed10882-f7ce-4f57-bb6b-b70b8eac5340-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp\" (UID: \"0ed10882-f7ce-4f57-bb6b-b70b8eac5340\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.798147 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.857455 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-6dd7dd855f-hw4fk"] Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.899317 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bde9e722-7f00-444e-92fe-ac0db62b6172-observability-operator-tls\") pod \"observability-operator-6dd7dd855f-hw4fk\" (UID: \"bde9e722-7f00-444e-92fe-ac0db62b6172\") " pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" Mar 17 15:50:33 crc kubenswrapper[4767]: I0317 15:50:33.899419 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh69m\" (UniqueName: \"kubernetes.io/projected/bde9e722-7f00-444e-92fe-ac0db62b6172-kube-api-access-rh69m\") pod \"observability-operator-6dd7dd855f-hw4fk\" (UID: \"bde9e722-7f00-444e-92fe-ac0db62b6172\") " pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.000856 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bde9e722-7f00-444e-92fe-ac0db62b6172-observability-operator-tls\") pod \"observability-operator-6dd7dd855f-hw4fk\" (UID: \"bde9e722-7f00-444e-92fe-ac0db62b6172\") " pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.001007 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh69m\" (UniqueName: \"kubernetes.io/projected/bde9e722-7f00-444e-92fe-ac0db62b6172-kube-api-access-rh69m\") pod \"observability-operator-6dd7dd855f-hw4fk\" (UID: \"bde9e722-7f00-444e-92fe-ac0db62b6172\") " pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.021246 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bde9e722-7f00-444e-92fe-ac0db62b6172-observability-operator-tls\") pod \"observability-operator-6dd7dd855f-hw4fk\" (UID: \"bde9e722-7f00-444e-92fe-ac0db62b6172\") " pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.177195 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.177317 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.284489 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh69m\" (UniqueName: \"kubernetes.io/projected/bde9e722-7f00-444e-92fe-ac0db62b6172-kube-api-access-rh69m\") pod \"observability-operator-6dd7dd855f-hw4fk\" (UID: \"bde9e722-7f00-444e-92fe-ac0db62b6172\") " pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.550655 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.625120 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t"] Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.779316 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp"] Mar 17 15:50:34 crc kubenswrapper[4767]: W0317 15:50:34.804348 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ed10882_f7ce_4f57_bb6b_b70b8eac5340.slice/crio-3f1cecc11c19daa377a6b133cb40d36fc14a1d7b84559c57ce1977dee911b382 WatchSource:0}: Error finding container 3f1cecc11c19daa377a6b133cb40d36fc14a1d7b84559c57ce1977dee911b382: Status 404 returned error can't find the container with id 3f1cecc11c19daa377a6b133cb40d36fc14a1d7b84559c57ce1977dee911b382 Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.891722 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-6d5d56fd99-rxwwn"] Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.892820 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.895655 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-service-cert" Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.896060 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-9crlw" Mar 17 15:50:34 crc kubenswrapper[4767]: I0317 15:50:34.923580 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-6d5d56fd99-rxwwn"] Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.016555 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2531b56b-9f38-4cef-9812-b1b30c29405a-apiservice-cert\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.016620 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2531b56b-9f38-4cef-9812-b1b30c29405a-webhook-cert\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.016640 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgh8j\" (UniqueName: \"kubernetes.io/projected/2531b56b-9f38-4cef-9812-b1b30c29405a-kube-api-access-hgh8j\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.016684 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2531b56b-9f38-4cef-9812-b1b30c29405a-openshift-service-ca\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.101692 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-6dd7dd855f-hw4fk"] Mar 17 15:50:35 crc kubenswrapper[4767]: W0317 15:50:35.104137 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbde9e722_7f00_444e_92fe_ac0db62b6172.slice/crio-4feb9344a03f1e421cc23826fbe2d0c343bb971c0a1890541200b38ff6cd832f WatchSource:0}: Error finding container 4feb9344a03f1e421cc23826fbe2d0c343bb971c0a1890541200b38ff6cd832f: Status 404 returned error can't find the container with id 4feb9344a03f1e421cc23826fbe2d0c343bb971c0a1890541200b38ff6cd832f Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.118708 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2531b56b-9f38-4cef-9812-b1b30c29405a-apiservice-cert\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.118778 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2531b56b-9f38-4cef-9812-b1b30c29405a-webhook-cert\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.118811 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgh8j\" (UniqueName: \"kubernetes.io/projected/2531b56b-9f38-4cef-9812-b1b30c29405a-kube-api-access-hgh8j\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.118875 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2531b56b-9f38-4cef-9812-b1b30c29405a-openshift-service-ca\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.121032 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2531b56b-9f38-4cef-9812-b1b30c29405a-openshift-service-ca\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.127304 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2531b56b-9f38-4cef-9812-b1b30c29405a-webhook-cert\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.127388 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2531b56b-9f38-4cef-9812-b1b30c29405a-apiservice-cert\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.143321 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgh8j\" (UniqueName: \"kubernetes.io/projected/2531b56b-9f38-4cef-9812-b1b30c29405a-kube-api-access-hgh8j\") pod \"perses-operator-6d5d56fd99-rxwwn\" (UID: \"2531b56b-9f38-4cef-9812-b1b30c29405a\") " pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.246578 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.624644 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" event={"ID":"bde9e722-7f00-444e-92fe-ac0db62b6172","Type":"ContainerStarted","Data":"4feb9344a03f1e421cc23826fbe2d0c343bb971c0a1890541200b38ff6cd832f"} Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.629328 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp" event={"ID":"0ed10882-f7ce-4f57-bb6b-b70b8eac5340","Type":"ContainerStarted","Data":"3f1cecc11c19daa377a6b133cb40d36fc14a1d7b84559c57ce1977dee911b382"} Mar 17 15:50:35 crc kubenswrapper[4767]: I0317 15:50:35.630506 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t" event={"ID":"aa728909-9d71-4e9c-a07c-e427969aa381","Type":"ContainerStarted","Data":"17e1de6fbb4fbbc2f7ce71420501519b4800dd4b9503382404d53340b8805466"} Mar 17 15:50:36 crc kubenswrapper[4767]: I0317 15:50:36.376098 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-6d5d56fd99-rxwwn"] Mar 17 15:50:36 crc kubenswrapper[4767]: I0317 15:50:36.688246 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" event={"ID":"2531b56b-9f38-4cef-9812-b1b30c29405a","Type":"ContainerStarted","Data":"41f622857ec55caf0aeda81d2635348172f8bbb5ecea386d2960291abcf25b41"} Mar 17 15:50:45 crc kubenswrapper[4767]: I0317 15:50:45.575742 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rh9xl" Mar 17 15:50:46 crc kubenswrapper[4767]: I0317 15:50:46.354555 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:46 crc kubenswrapper[4767]: I0317 15:50:46.356012 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" Mar 17 15:50:53 crc kubenswrapper[4767]: E0317 15:50:53.972535 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:f78b160ba3b815f53d6a72425f3f3a9d7946795177bd68c7c614fa84f97be630" Mar 17 15:50:53 crc kubenswrapper[4767]: E0317 15:50:53.975091 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:f78b160ba3b815f53d6a72425f3f3a9d7946795177bd68c7c614fa84f97be630,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.4.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hgh8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-6d5d56fd99-rxwwn_openshift-operators(2531b56b-9f38-4cef-9812-b1b30c29405a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 15:50:53 crc kubenswrapper[4767]: E0317 15:50:53.976724 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" podUID="2531b56b-9f38-4cef-9812-b1b30c29405a" Mar 17 15:50:54 crc kubenswrapper[4767]: E0317 15:50:54.316429 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:f78b160ba3b815f53d6a72425f3f3a9d7946795177bd68c7c614fa84f97be630\\\"\"" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" podUID="2531b56b-9f38-4cef-9812-b1b30c29405a" Mar 17 15:50:54 crc kubenswrapper[4767]: I0317 15:50:54.427058 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6"] Mar 17 15:50:55 crc kubenswrapper[4767]: I0317 15:50:55.319435 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" event={"ID":"bde9e722-7f00-444e-92fe-ac0db62b6172","Type":"ContainerStarted","Data":"6eb79470cca80080f20b19c5b3889f9f84c4848d082836033447d6cca7772525"} Mar 17 15:50:55 crc kubenswrapper[4767]: I0317 15:50:55.321599 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" event={"ID":"1373b043-07be-4e46-9913-4ba92b42352b","Type":"ContainerStarted","Data":"89f9fb5f25ffe8c716e91f41785a11de43e42972ae46e9e66417f6faf76cb228"} Mar 17 15:50:55 crc kubenswrapper[4767]: I0317 15:50:55.321730 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" Mar 17 15:50:55 crc kubenswrapper[4767]: I0317 15:50:55.323742 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp" event={"ID":"0ed10882-f7ce-4f57-bb6b-b70b8eac5340","Type":"ContainerStarted","Data":"095518840e659597163d5542e7d92cd8469bccb21f7cc099270a1db7687b2423"} Mar 17 15:50:55 crc kubenswrapper[4767]: I0317 15:50:55.328132 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t" event={"ID":"aa728909-9d71-4e9c-a07c-e427969aa381","Type":"ContainerStarted","Data":"6c8e1744ca24c3baa9ab692931657860eb9e6efdd38b8ef3570c6faeaa75d331"} Mar 17 15:50:55 crc kubenswrapper[4767]: I0317 15:50:55.347358 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" podStartSLOduration=3.398938902 podStartE2EDuration="22.347327217s" podCreationTimestamp="2026-03-17 15:50:33 +0000 UTC" firstStartedPulling="2026-03-17 15:50:35.107645228 +0000 UTC m=+826.520961275" lastFinishedPulling="2026-03-17 15:50:54.056033543 +0000 UTC m=+845.469349590" observedRunningTime="2026-03-17 15:50:55.343109784 +0000 UTC m=+846.756425841" watchObservedRunningTime="2026-03-17 15:50:55.347327217 +0000 UTC m=+846.760643264" Mar 17 15:50:55 crc kubenswrapper[4767]: I0317 15:50:55.387379 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-ms4xp" podStartSLOduration=3.142793213 podStartE2EDuration="22.387347247s" podCreationTimestamp="2026-03-17 15:50:33 +0000 UTC" firstStartedPulling="2026-03-17 15:50:34.81151613 +0000 UTC m=+826.224832177" lastFinishedPulling="2026-03-17 15:50:54.056070164 +0000 UTC m=+845.469386211" observedRunningTime="2026-03-17 15:50:55.381985381 +0000 UTC m=+846.795301428" watchObservedRunningTime="2026-03-17 15:50:55.387347247 +0000 UTC m=+846.800663294" Mar 17 15:50:55 crc kubenswrapper[4767]: I0317 15:50:55.419318 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b547d67d-4kj4t" podStartSLOduration=3.041697058 podStartE2EDuration="22.419287921s" podCreationTimestamp="2026-03-17 15:50:33 +0000 UTC" firstStartedPulling="2026-03-17 15:50:34.656197479 +0000 UTC m=+826.069513526" lastFinishedPulling="2026-03-17 15:50:54.033788342 +0000 UTC m=+845.447104389" observedRunningTime="2026-03-17 15:50:55.41652299 +0000 UTC m=+846.829839047" watchObservedRunningTime="2026-03-17 15:50:55.419287921 +0000 UTC m=+846.832603968" Mar 17 15:50:56 crc kubenswrapper[4767]: I0317 15:50:56.361520 4767 patch_prober.go:28] interesting pod/observability-operator-6dd7dd855f-hw4fk container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:50:56 crc kubenswrapper[4767]: I0317 15:50:56.361630 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" podUID="bde9e722-7f00-444e-92fe-ac0db62b6172" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 15:50:56 crc kubenswrapper[4767]: I0317 15:50:56.458249 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" Mar 17 15:50:59 crc kubenswrapper[4767]: I0317 15:50:59.374699 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" event={"ID":"1373b043-07be-4e46-9913-4ba92b42352b","Type":"ContainerStarted","Data":"0331cfa03a6d6a2fe0dec0bd7f3e58732729c71f50870995044ad946ac40bd4d"} Mar 17 15:50:59 crc kubenswrapper[4767]: I0317 15:50:59.408927 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-8ff7d675-j8nb6" podStartSLOduration=23.408436554 podStartE2EDuration="27.408901368s" podCreationTimestamp="2026-03-17 15:50:32 +0000 UTC" firstStartedPulling="2026-03-17 15:50:54.456027288 +0000 UTC m=+845.869343335" lastFinishedPulling="2026-03-17 15:50:58.456492102 +0000 UTC m=+849.869808149" observedRunningTime="2026-03-17 15:50:59.405948252 +0000 UTC m=+850.819264309" watchObservedRunningTime="2026-03-17 15:50:59.408901368 +0000 UTC m=+850.822217415" Mar 17 15:51:04 crc kubenswrapper[4767]: I0317 15:51:04.166885 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:51:04 crc kubenswrapper[4767]: I0317 15:51:04.167929 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.305617 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-p8bj4"] Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.308207 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-p8bj4" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.312487 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.312849 4767 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-ftj4b" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.313606 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.329839 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-z7fsr"] Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.331026 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-z7fsr" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.339254 4767 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-z2c8p" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.345755 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-p8bj4"] Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.398903 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-kv9bb"] Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.400990 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.405099 4767 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-dld97" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.414360 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvq97\" (UniqueName: \"kubernetes.io/projected/804a73e2-b63c-4968-b028-7d7eb18b5f3d-kube-api-access-xvq97\") pod \"cert-manager-cainjector-cf98fcc89-p8bj4\" (UID: \"804a73e2-b63c-4968-b028-7d7eb18b5f3d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-p8bj4" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.419781 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-z7fsr"] Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.432046 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-kv9bb"] Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.521838 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5dwl\" (UniqueName: \"kubernetes.io/projected/baf4c43e-08e2-4afa-b9b5-8bb4aa8f0352-kube-api-access-t5dwl\") pod \"cert-manager-webhook-687f57d79b-kv9bb\" (UID: \"baf4c43e-08e2-4afa-b9b5-8bb4aa8f0352\") " pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.521939 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvq97\" (UniqueName: \"kubernetes.io/projected/804a73e2-b63c-4968-b028-7d7eb18b5f3d-kube-api-access-xvq97\") pod \"cert-manager-cainjector-cf98fcc89-p8bj4\" (UID: \"804a73e2-b63c-4968-b028-7d7eb18b5f3d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-p8bj4" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.521978 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrwjc\" (UniqueName: \"kubernetes.io/projected/13bb25a7-a93f-44d5-9d8f-6d4ecbe0a945-kube-api-access-wrwjc\") pod \"cert-manager-858654f9db-z7fsr\" (UID: \"13bb25a7-a93f-44d5-9d8f-6d4ecbe0a945\") " pod="cert-manager/cert-manager-858654f9db-z7fsr" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.543512 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvq97\" (UniqueName: \"kubernetes.io/projected/804a73e2-b63c-4968-b028-7d7eb18b5f3d-kube-api-access-xvq97\") pod \"cert-manager-cainjector-cf98fcc89-p8bj4\" (UID: \"804a73e2-b63c-4968-b028-7d7eb18b5f3d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-p8bj4" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.623429 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5dwl\" (UniqueName: \"kubernetes.io/projected/baf4c43e-08e2-4afa-b9b5-8bb4aa8f0352-kube-api-access-t5dwl\") pod \"cert-manager-webhook-687f57d79b-kv9bb\" (UID: \"baf4c43e-08e2-4afa-b9b5-8bb4aa8f0352\") " pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.623576 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrwjc\" (UniqueName: \"kubernetes.io/projected/13bb25a7-a93f-44d5-9d8f-6d4ecbe0a945-kube-api-access-wrwjc\") pod \"cert-manager-858654f9db-z7fsr\" (UID: \"13bb25a7-a93f-44d5-9d8f-6d4ecbe0a945\") " pod="cert-manager/cert-manager-858654f9db-z7fsr" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.641972 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-p8bj4" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.656280 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5dwl\" (UniqueName: \"kubernetes.io/projected/baf4c43e-08e2-4afa-b9b5-8bb4aa8f0352-kube-api-access-t5dwl\") pod \"cert-manager-webhook-687f57d79b-kv9bb\" (UID: \"baf4c43e-08e2-4afa-b9b5-8bb4aa8f0352\") " pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.659791 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrwjc\" (UniqueName: \"kubernetes.io/projected/13bb25a7-a93f-44d5-9d8f-6d4ecbe0a945-kube-api-access-wrwjc\") pod \"cert-manager-858654f9db-z7fsr\" (UID: \"13bb25a7-a93f-44d5-9d8f-6d4ecbe0a945\") " pod="cert-manager/cert-manager-858654f9db-z7fsr" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.690520 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-z7fsr" Mar 17 15:51:07 crc kubenswrapper[4767]: I0317 15:51:07.724829 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" Mar 17 15:51:08 crc kubenswrapper[4767]: I0317 15:51:08.176164 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-z7fsr"] Mar 17 15:51:08 crc kubenswrapper[4767]: I0317 15:51:08.193756 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 15:51:08 crc kubenswrapper[4767]: I0317 15:51:08.222841 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-kv9bb"] Mar 17 15:51:08 crc kubenswrapper[4767]: I0317 15:51:08.291032 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-p8bj4"] Mar 17 15:51:08 crc kubenswrapper[4767]: W0317 15:51:08.292586 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod804a73e2_b63c_4968_b028_7d7eb18b5f3d.slice/crio-432333a255a8b49e9b93d4ab86ad8ea66749083716f8c626b00227c817533299 WatchSource:0}: Error finding container 432333a255a8b49e9b93d4ab86ad8ea66749083716f8c626b00227c817533299: Status 404 returned error can't find the container with id 432333a255a8b49e9b93d4ab86ad8ea66749083716f8c626b00227c817533299 Mar 17 15:51:08 crc kubenswrapper[4767]: I0317 15:51:08.449768 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" event={"ID":"baf4c43e-08e2-4afa-b9b5-8bb4aa8f0352","Type":"ContainerStarted","Data":"f2a7733ec9fd03d2f81481d9d57d1542e20456cf4bd190ae7f2426f9c5084461"} Mar 17 15:51:08 crc kubenswrapper[4767]: I0317 15:51:08.450738 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-z7fsr" event={"ID":"13bb25a7-a93f-44d5-9d8f-6d4ecbe0a945","Type":"ContainerStarted","Data":"3bd88998fb7a050bfa00e2476aa3a7d7f3d972970c97d021d6ea3b25faeee6f3"} Mar 17 15:51:08 crc kubenswrapper[4767]: I0317 15:51:08.451849 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-p8bj4" event={"ID":"804a73e2-b63c-4968-b028-7d7eb18b5f3d","Type":"ContainerStarted","Data":"432333a255a8b49e9b93d4ab86ad8ea66749083716f8c626b00227c817533299"} Mar 17 15:51:12 crc kubenswrapper[4767]: I0317 15:51:12.491829 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" event={"ID":"2531b56b-9f38-4cef-9812-b1b30c29405a","Type":"ContainerStarted","Data":"80750dc5612eb0b6115c32d41a738ddceb5c7851a26a737816621ccd5a9796ab"} Mar 17 15:51:12 crc kubenswrapper[4767]: I0317 15:51:12.497050 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:51:12 crc kubenswrapper[4767]: I0317 15:51:12.529378 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" podStartSLOduration=3.392927533 podStartE2EDuration="38.529203834s" podCreationTimestamp="2026-03-17 15:50:34 +0000 UTC" firstStartedPulling="2026-03-17 15:50:36.460614655 +0000 UTC m=+827.873930702" lastFinishedPulling="2026-03-17 15:51:11.596890956 +0000 UTC m=+863.010207003" observedRunningTime="2026-03-17 15:51:12.520455988 +0000 UTC m=+863.933772035" watchObservedRunningTime="2026-03-17 15:51:12.529203834 +0000 UTC m=+863.942519901" Mar 17 15:51:16 crc kubenswrapper[4767]: I0317 15:51:16.525012 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-p8bj4" event={"ID":"804a73e2-b63c-4968-b028-7d7eb18b5f3d","Type":"ContainerStarted","Data":"5028c7a8e52a0668db50dd8bbb1232c9def5df15eb9f059ea3671ccc5decc4b6"} Mar 17 15:51:16 crc kubenswrapper[4767]: I0317 15:51:16.529332 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" event={"ID":"baf4c43e-08e2-4afa-b9b5-8bb4aa8f0352","Type":"ContainerStarted","Data":"21f9afae7648c23c4f8929c9f1995dc20d4322bf8faaed8db8a09dbcf8516a59"} Mar 17 15:51:16 crc kubenswrapper[4767]: I0317 15:51:16.531216 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-z7fsr" event={"ID":"13bb25a7-a93f-44d5-9d8f-6d4ecbe0a945","Type":"ContainerStarted","Data":"efea5d0a21e233ee1f3603115810ac0612b72918fb9d820ee28a3587884c520e"} Mar 17 15:51:16 crc kubenswrapper[4767]: I0317 15:51:16.550040 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-p8bj4" podStartSLOduration=2.6559608470000002 podStartE2EDuration="9.550003452s" podCreationTimestamp="2026-03-17 15:51:07 +0000 UTC" firstStartedPulling="2026-03-17 15:51:08.295676816 +0000 UTC m=+859.708992863" lastFinishedPulling="2026-03-17 15:51:15.189719421 +0000 UTC m=+866.603035468" observedRunningTime="2026-03-17 15:51:16.544660076 +0000 UTC m=+867.957976153" watchObservedRunningTime="2026-03-17 15:51:16.550003452 +0000 UTC m=+867.963319529" Mar 17 15:51:16 crc kubenswrapper[4767]: I0317 15:51:16.597029 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" podStartSLOduration=2.620391506 podStartE2EDuration="9.596994286s" podCreationTimestamp="2026-03-17 15:51:07 +0000 UTC" firstStartedPulling="2026-03-17 15:51:08.234915659 +0000 UTC m=+859.648231706" lastFinishedPulling="2026-03-17 15:51:15.211518439 +0000 UTC m=+866.624834486" observedRunningTime="2026-03-17 15:51:16.571221372 +0000 UTC m=+867.984537439" watchObservedRunningTime="2026-03-17 15:51:16.596994286 +0000 UTC m=+868.010310343" Mar 17 15:51:16 crc kubenswrapper[4767]: I0317 15:51:16.598336 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-z7fsr" podStartSLOduration=2.579018357 podStartE2EDuration="9.598328275s" podCreationTimestamp="2026-03-17 15:51:07 +0000 UTC" firstStartedPulling="2026-03-17 15:51:08.193389495 +0000 UTC m=+859.606705542" lastFinishedPulling="2026-03-17 15:51:15.212699413 +0000 UTC m=+866.626015460" observedRunningTime="2026-03-17 15:51:16.59611187 +0000 UTC m=+868.009427927" watchObservedRunningTime="2026-03-17 15:51:16.598328275 +0000 UTC m=+868.011644332" Mar 17 15:51:17 crc kubenswrapper[4767]: I0317 15:51:17.539288 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" Mar 17 15:51:22 crc kubenswrapper[4767]: I0317 15:51:22.729150 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" Mar 17 15:51:25 crc kubenswrapper[4767]: I0317 15:51:25.250696 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" Mar 17 15:51:34 crc kubenswrapper[4767]: I0317 15:51:34.166662 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:51:34 crc kubenswrapper[4767]: I0317 15:51:34.167608 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:51:34 crc kubenswrapper[4767]: I0317 15:51:34.167688 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:51:34 crc kubenswrapper[4767]: I0317 15:51:34.168588 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"942ec751e9132c97b85320d70dbc413daf5abb6e5090f70bbebaf738b7e2ef34"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 15:51:34 crc kubenswrapper[4767]: I0317 15:51:34.168651 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://942ec751e9132c97b85320d70dbc413daf5abb6e5090f70bbebaf738b7e2ef34" gracePeriod=600 Mar 17 15:51:34 crc kubenswrapper[4767]: I0317 15:51:34.665755 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="942ec751e9132c97b85320d70dbc413daf5abb6e5090f70bbebaf738b7e2ef34" exitCode=0 Mar 17 15:51:34 crc kubenswrapper[4767]: I0317 15:51:34.666357 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"942ec751e9132c97b85320d70dbc413daf5abb6e5090f70bbebaf738b7e2ef34"} Mar 17 15:51:34 crc kubenswrapper[4767]: I0317 15:51:34.666401 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"265b2d9bdc4eb65af73c1e843186a34dfa697843aaf572fda7ed43575c18359a"} Mar 17 15:51:34 crc kubenswrapper[4767]: I0317 15:51:34.666423 4767 scope.go:117] "RemoveContainer" containerID="d152bc4b2f69796edf7aec43d7f21284777821024a613d49605f7c9d9db18719" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.188418 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl"] Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.190902 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.193440 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.207270 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl"] Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.380904 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-util\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.381797 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-bundle\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.382026 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2dh9\" (UniqueName: \"kubernetes.io/projected/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-kube-api-access-r2dh9\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.386712 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl"] Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.388265 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.441232 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl"] Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.484087 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-util\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.484182 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-bundle\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.484800 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2dh9\" (UniqueName: \"kubernetes.io/projected/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-kube-api-access-r2dh9\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.485006 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-util\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.485088 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-bundle\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.518325 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2dh9\" (UniqueName: \"kubernetes.io/projected/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-kube-api-access-r2dh9\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.586629 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-util\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.587112 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-bundle\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.587292 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fmc6\" (UniqueName: \"kubernetes.io/projected/12d0e8e6-6998-4a32-b1c3-7203f88a534f-kube-api-access-9fmc6\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.689113 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fmc6\" (UniqueName: \"kubernetes.io/projected/12d0e8e6-6998-4a32-b1c3-7203f88a534f-kube-api-access-9fmc6\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.689211 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-util\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.689261 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-bundle\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.690613 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-util\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.690646 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-bundle\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.707578 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fmc6\" (UniqueName: \"kubernetes.io/projected/12d0e8e6-6998-4a32-b1c3-7203f88a534f-kube-api-access-9fmc6\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:51 crc kubenswrapper[4767]: I0317 15:51:51.817042 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:52 crc kubenswrapper[4767]: I0317 15:51:52.008785 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:52 crc kubenswrapper[4767]: I0317 15:51:52.189271 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl"] Mar 17 15:51:52 crc kubenswrapper[4767]: W0317 15:51:52.201324 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a89f5d9_5849_436a_9c51_d4f6a4f1aaed.slice/crio-212a9fda2dca0dbf847a85181c104db81161180d6cd4e82e499b915815c68077 WatchSource:0}: Error finding container 212a9fda2dca0dbf847a85181c104db81161180d6cd4e82e499b915815c68077: Status 404 returned error can't find the container with id 212a9fda2dca0dbf847a85181c104db81161180d6cd4e82e499b915815c68077 Mar 17 15:51:52 crc kubenswrapper[4767]: I0317 15:51:52.355976 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl"] Mar 17 15:51:52 crc kubenswrapper[4767]: W0317 15:51:52.366979 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12d0e8e6_6998_4a32_b1c3_7203f88a534f.slice/crio-9744f127262f5f2c8b40b852f4675098b15e04f410e1804f49310c9431313a04 WatchSource:0}: Error finding container 9744f127262f5f2c8b40b852f4675098b15e04f410e1804f49310c9431313a04: Status 404 returned error can't find the container with id 9744f127262f5f2c8b40b852f4675098b15e04f410e1804f49310c9431313a04 Mar 17 15:51:52 crc kubenswrapper[4767]: I0317 15:51:52.816752 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" event={"ID":"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed","Type":"ContainerStarted","Data":"f44945f2d110f488aeded2da02469574f787e1d0cf0487ff30f57457a5a3881d"} Mar 17 15:51:52 crc kubenswrapper[4767]: I0317 15:51:52.816817 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" event={"ID":"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed","Type":"ContainerStarted","Data":"212a9fda2dca0dbf847a85181c104db81161180d6cd4e82e499b915815c68077"} Mar 17 15:51:52 crc kubenswrapper[4767]: I0317 15:51:52.819015 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" event={"ID":"12d0e8e6-6998-4a32-b1c3-7203f88a534f","Type":"ContainerStarted","Data":"f7b92a512725a35864f856d9f7a3332316f6e72723c3f231c91e612aaf3b515c"} Mar 17 15:51:52 crc kubenswrapper[4767]: I0317 15:51:52.819072 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" event={"ID":"12d0e8e6-6998-4a32-b1c3-7203f88a534f","Type":"ContainerStarted","Data":"9744f127262f5f2c8b40b852f4675098b15e04f410e1804f49310c9431313a04"} Mar 17 15:51:53 crc kubenswrapper[4767]: I0317 15:51:53.830634 4767 generic.go:334] "Generic (PLEG): container finished" podID="12d0e8e6-6998-4a32-b1c3-7203f88a534f" containerID="f7b92a512725a35864f856d9f7a3332316f6e72723c3f231c91e612aaf3b515c" exitCode=0 Mar 17 15:51:53 crc kubenswrapper[4767]: I0317 15:51:53.830753 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" event={"ID":"12d0e8e6-6998-4a32-b1c3-7203f88a534f","Type":"ContainerDied","Data":"f7b92a512725a35864f856d9f7a3332316f6e72723c3f231c91e612aaf3b515c"} Mar 17 15:51:53 crc kubenswrapper[4767]: I0317 15:51:53.834688 4767 generic.go:334] "Generic (PLEG): container finished" podID="2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" containerID="f44945f2d110f488aeded2da02469574f787e1d0cf0487ff30f57457a5a3881d" exitCode=0 Mar 17 15:51:53 crc kubenswrapper[4767]: I0317 15:51:53.834730 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" event={"ID":"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed","Type":"ContainerDied","Data":"f44945f2d110f488aeded2da02469574f787e1d0cf0487ff30f57457a5a3881d"} Mar 17 15:51:54 crc kubenswrapper[4767]: I0317 15:51:54.924086 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-msl8c"] Mar 17 15:51:54 crc kubenswrapper[4767]: I0317 15:51:54.927746 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:51:54 crc kubenswrapper[4767]: I0317 15:51:54.939836 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-msl8c"] Mar 17 15:51:54 crc kubenswrapper[4767]: I0317 15:51:54.957444 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-catalog-content\") pod \"redhat-operators-msl8c\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:51:54 crc kubenswrapper[4767]: I0317 15:51:54.957605 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-utilities\") pod \"redhat-operators-msl8c\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:51:54 crc kubenswrapper[4767]: I0317 15:51:54.957647 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57htg\" (UniqueName: \"kubernetes.io/projected/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-kube-api-access-57htg\") pod \"redhat-operators-msl8c\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:51:55 crc kubenswrapper[4767]: I0317 15:51:55.214541 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-utilities\") pod \"redhat-operators-msl8c\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:51:55 crc kubenswrapper[4767]: I0317 15:51:55.214630 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57htg\" (UniqueName: \"kubernetes.io/projected/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-kube-api-access-57htg\") pod \"redhat-operators-msl8c\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:51:55 crc kubenswrapper[4767]: I0317 15:51:55.214720 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-catalog-content\") pod \"redhat-operators-msl8c\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:51:55 crc kubenswrapper[4767]: I0317 15:51:55.215569 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-catalog-content\") pod \"redhat-operators-msl8c\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:51:55 crc kubenswrapper[4767]: I0317 15:51:55.215934 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-utilities\") pod \"redhat-operators-msl8c\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:51:55 crc kubenswrapper[4767]: I0317 15:51:55.329560 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57htg\" (UniqueName: \"kubernetes.io/projected/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-kube-api-access-57htg\") pod \"redhat-operators-msl8c\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:51:55 crc kubenswrapper[4767]: I0317 15:51:55.549341 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:51:55 crc kubenswrapper[4767]: I0317 15:51:55.852869 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" event={"ID":"12d0e8e6-6998-4a32-b1c3-7203f88a534f","Type":"ContainerStarted","Data":"4fe7d40c946b1461abbf2fd4b87acc44937e37a5fa4cd40ec75fa2e13a188576"} Mar 17 15:51:56 crc kubenswrapper[4767]: I0317 15:51:56.010726 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-msl8c"] Mar 17 15:51:56 crc kubenswrapper[4767]: W0317 15:51:56.084009 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ae678ad_31d9_4eb5_84ba_ab0fbd24313b.slice/crio-100e6816cb0f33163699474491989651207e22422601307fc02ce93f66fd980e WatchSource:0}: Error finding container 100e6816cb0f33163699474491989651207e22422601307fc02ce93f66fd980e: Status 404 returned error can't find the container with id 100e6816cb0f33163699474491989651207e22422601307fc02ce93f66fd980e Mar 17 15:51:56 crc kubenswrapper[4767]: I0317 15:51:56.864786 4767 generic.go:334] "Generic (PLEG): container finished" podID="2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" containerID="1420da57fb735f8dbb0f18284148b012a99fa55f0813b4d1f457d28e134fd162" exitCode=0 Mar 17 15:51:56 crc kubenswrapper[4767]: I0317 15:51:56.864926 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" event={"ID":"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed","Type":"ContainerDied","Data":"1420da57fb735f8dbb0f18284148b012a99fa55f0813b4d1f457d28e134fd162"} Mar 17 15:51:56 crc kubenswrapper[4767]: I0317 15:51:56.868084 4767 generic.go:334] "Generic (PLEG): container finished" podID="12d0e8e6-6998-4a32-b1c3-7203f88a534f" containerID="4fe7d40c946b1461abbf2fd4b87acc44937e37a5fa4cd40ec75fa2e13a188576" exitCode=0 Mar 17 15:51:56 crc kubenswrapper[4767]: I0317 15:51:56.868220 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" event={"ID":"12d0e8e6-6998-4a32-b1c3-7203f88a534f","Type":"ContainerDied","Data":"4fe7d40c946b1461abbf2fd4b87acc44937e37a5fa4cd40ec75fa2e13a188576"} Mar 17 15:51:56 crc kubenswrapper[4767]: I0317 15:51:56.871790 4767 generic.go:334] "Generic (PLEG): container finished" podID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerID="a21f8cec437b64ce1e106119a6a73e87fc9653ca5cf8833a0907b7b9f1635c4d" exitCode=0 Mar 17 15:51:56 crc kubenswrapper[4767]: I0317 15:51:56.871843 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msl8c" event={"ID":"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b","Type":"ContainerDied","Data":"a21f8cec437b64ce1e106119a6a73e87fc9653ca5cf8833a0907b7b9f1635c4d"} Mar 17 15:51:56 crc kubenswrapper[4767]: I0317 15:51:56.871874 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msl8c" event={"ID":"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b","Type":"ContainerStarted","Data":"100e6816cb0f33163699474491989651207e22422601307fc02ce93f66fd980e"} Mar 17 15:51:57 crc kubenswrapper[4767]: I0317 15:51:57.882466 4767 generic.go:334] "Generic (PLEG): container finished" podID="12d0e8e6-6998-4a32-b1c3-7203f88a534f" containerID="67b29947b2b6557d151629cebd331e284b1bd9b4160ff65bdccd035e2dbeb42c" exitCode=0 Mar 17 15:51:57 crc kubenswrapper[4767]: I0317 15:51:57.882587 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" event={"ID":"12d0e8e6-6998-4a32-b1c3-7203f88a534f","Type":"ContainerDied","Data":"67b29947b2b6557d151629cebd331e284b1bd9b4160ff65bdccd035e2dbeb42c"} Mar 17 15:51:57 crc kubenswrapper[4767]: I0317 15:51:57.884650 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msl8c" event={"ID":"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b","Type":"ContainerStarted","Data":"eb96fe5161d6c9c53ec50e1e3626bbf6d862d764960a33730a3571ef82cbeef2"} Mar 17 15:51:57 crc kubenswrapper[4767]: I0317 15:51:57.886991 4767 generic.go:334] "Generic (PLEG): container finished" podID="2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" containerID="b19f8a6459884b16e7733d227791444b4d2878a0c2300e7b51d7378301793992" exitCode=0 Mar 17 15:51:57 crc kubenswrapper[4767]: I0317 15:51:57.887032 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" event={"ID":"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed","Type":"ContainerDied","Data":"b19f8a6459884b16e7733d227791444b4d2878a0c2300e7b51d7378301793992"} Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.386116 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.523909 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-util\") pod \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.523974 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.524121 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fmc6\" (UniqueName: \"kubernetes.io/projected/12d0e8e6-6998-4a32-b1c3-7203f88a534f-kube-api-access-9fmc6\") pod \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.524250 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-bundle\") pod \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\" (UID: \"12d0e8e6-6998-4a32-b1c3-7203f88a534f\") " Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.525716 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-bundle" (OuterVolumeSpecName: "bundle") pod "12d0e8e6-6998-4a32-b1c3-7203f88a534f" (UID: "12d0e8e6-6998-4a32-b1c3-7203f88a534f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.535398 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12d0e8e6-6998-4a32-b1c3-7203f88a534f-kube-api-access-9fmc6" (OuterVolumeSpecName: "kube-api-access-9fmc6") pod "12d0e8e6-6998-4a32-b1c3-7203f88a534f" (UID: "12d0e8e6-6998-4a32-b1c3-7203f88a534f"). InnerVolumeSpecName "kube-api-access-9fmc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.537255 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-util" (OuterVolumeSpecName: "util") pod "12d0e8e6-6998-4a32-b1c3-7203f88a534f" (UID: "12d0e8e6-6998-4a32-b1c3-7203f88a534f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.626574 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-util\") pod \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.626706 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2dh9\" (UniqueName: \"kubernetes.io/projected/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-kube-api-access-r2dh9\") pod \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.626843 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-bundle\") pod \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\" (UID: \"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed\") " Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.627117 4767 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-util\") on node \"crc\" DevicePath \"\"" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.627156 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fmc6\" (UniqueName: \"kubernetes.io/projected/12d0e8e6-6998-4a32-b1c3-7203f88a534f-kube-api-access-9fmc6\") on node \"crc\" DevicePath \"\"" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.627187 4767 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12d0e8e6-6998-4a32-b1c3-7203f88a534f-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.628274 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-bundle" (OuterVolumeSpecName: "bundle") pod "2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" (UID: "2a89f5d9-5849-436a-9c51-d4f6a4f1aaed"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.631458 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-kube-api-access-r2dh9" (OuterVolumeSpecName: "kube-api-access-r2dh9") pod "2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" (UID: "2a89f5d9-5849-436a-9c51-d4f6a4f1aaed"). InnerVolumeSpecName "kube-api-access-r2dh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.640194 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-util" (OuterVolumeSpecName: "util") pod "2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" (UID: "2a89f5d9-5849-436a-9c51-d4f6a4f1aaed"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.728505 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2dh9\" (UniqueName: \"kubernetes.io/projected/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-kube-api-access-r2dh9\") on node \"crc\" DevicePath \"\"" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.728548 4767 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:51:59 crc kubenswrapper[4767]: I0317 15:51:59.728565 4767 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2a89f5d9-5849-436a-9c51-d4f6a4f1aaed-util\") on node \"crc\" DevicePath \"\"" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.372057 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562712-r589g"] Mar 17 15:52:00 crc kubenswrapper[4767]: E0317 15:52:00.377369 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" containerName="extract" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.377407 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" containerName="extract" Mar 17 15:52:00 crc kubenswrapper[4767]: E0317 15:52:00.377435 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" containerName="pull" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.377443 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" containerName="pull" Mar 17 15:52:00 crc kubenswrapper[4767]: E0317 15:52:00.377506 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" containerName="util" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.377527 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" containerName="util" Mar 17 15:52:00 crc kubenswrapper[4767]: E0317 15:52:00.377558 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d0e8e6-6998-4a32-b1c3-7203f88a534f" containerName="util" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.377575 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d0e8e6-6998-4a32-b1c3-7203f88a534f" containerName="util" Mar 17 15:52:00 crc kubenswrapper[4767]: E0317 15:52:00.377616 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d0e8e6-6998-4a32-b1c3-7203f88a534f" containerName="extract" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.377634 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d0e8e6-6998-4a32-b1c3-7203f88a534f" containerName="extract" Mar 17 15:52:00 crc kubenswrapper[4767]: E0317 15:52:00.377661 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d0e8e6-6998-4a32-b1c3-7203f88a534f" containerName="pull" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.377673 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d0e8e6-6998-4a32-b1c3-7203f88a534f" containerName="pull" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.378236 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d0e8e6-6998-4a32-b1c3-7203f88a534f" containerName="extract" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.378284 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a89f5d9-5849-436a-9c51-d4f6a4f1aaed" containerName="extract" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.379285 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562712-r589g" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.384677 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.386621 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" event={"ID":"2a89f5d9-5849-436a-9c51-d4f6a4f1aaed","Type":"ContainerDied","Data":"212a9fda2dca0dbf847a85181c104db81161180d6cd4e82e499b915815c68077"} Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.386670 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="212a9fda2dca0dbf847a85181c104db81161180d6cd4e82e499b915815c68077" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.386892 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cbqzbl" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.387501 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.390035 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.399539 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562712-r589g"] Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.414190 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" event={"ID":"12d0e8e6-6998-4a32-b1c3-7203f88a534f","Type":"ContainerDied","Data":"9744f127262f5f2c8b40b852f4675098b15e04f410e1804f49310c9431313a04"} Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.414268 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9744f127262f5f2c8b40b852f4675098b15e04f410e1804f49310c9431313a04" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.414402 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d528lpl" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.505630 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4hbn\" (UniqueName: \"kubernetes.io/projected/f4ed68ee-0956-4d37-9d3b-f1904ff19451-kube-api-access-k4hbn\") pod \"auto-csr-approver-29562712-r589g\" (UID: \"f4ed68ee-0956-4d37-9d3b-f1904ff19451\") " pod="openshift-infra/auto-csr-approver-29562712-r589g" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.608857 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4hbn\" (UniqueName: \"kubernetes.io/projected/f4ed68ee-0956-4d37-9d3b-f1904ff19451-kube-api-access-k4hbn\") pod \"auto-csr-approver-29562712-r589g\" (UID: \"f4ed68ee-0956-4d37-9d3b-f1904ff19451\") " pod="openshift-infra/auto-csr-approver-29562712-r589g" Mar 17 15:52:00 crc kubenswrapper[4767]: I0317 15:52:00.749745 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4hbn\" (UniqueName: \"kubernetes.io/projected/f4ed68ee-0956-4d37-9d3b-f1904ff19451-kube-api-access-k4hbn\") pod \"auto-csr-approver-29562712-r589g\" (UID: \"f4ed68ee-0956-4d37-9d3b-f1904ff19451\") " pod="openshift-infra/auto-csr-approver-29562712-r589g" Mar 17 15:52:01 crc kubenswrapper[4767]: I0317 15:52:01.023764 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562712-r589g" Mar 17 15:52:01 crc kubenswrapper[4767]: I0317 15:52:01.567899 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562712-r589g"] Mar 17 15:52:02 crc kubenswrapper[4767]: I0317 15:52:02.443897 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562712-r589g" event={"ID":"f4ed68ee-0956-4d37-9d3b-f1904ff19451","Type":"ContainerStarted","Data":"5da8c9466851405b6631cb9b499584cce3f1f801525aee4864ee89ac6214196d"} Mar 17 15:52:02 crc kubenswrapper[4767]: I0317 15:52:02.446599 4767 generic.go:334] "Generic (PLEG): container finished" podID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerID="eb96fe5161d6c9c53ec50e1e3626bbf6d862d764960a33730a3571ef82cbeef2" exitCode=0 Mar 17 15:52:02 crc kubenswrapper[4767]: I0317 15:52:02.446656 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msl8c" event={"ID":"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b","Type":"ContainerDied","Data":"eb96fe5161d6c9c53ec50e1e3626bbf6d862d764960a33730a3571ef82cbeef2"} Mar 17 15:52:03 crc kubenswrapper[4767]: I0317 15:52:03.457781 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562712-r589g" event={"ID":"f4ed68ee-0956-4d37-9d3b-f1904ff19451","Type":"ContainerStarted","Data":"63135d6512e4bb5413b9841549fb5ec1a52f9e8411d5691cb6707e9465c12312"} Mar 17 15:52:03 crc kubenswrapper[4767]: I0317 15:52:03.464004 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msl8c" event={"ID":"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b","Type":"ContainerStarted","Data":"0b5d543001dab9e12a0aa329c64fbc38b3a1ef67afebfbe7607ec1f78683dd54"} Mar 17 15:52:03 crc kubenswrapper[4767]: I0317 15:52:03.516395 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562712-r589g" podStartSLOduration=2.4385518 podStartE2EDuration="3.516378992s" podCreationTimestamp="2026-03-17 15:52:00 +0000 UTC" firstStartedPulling="2026-03-17 15:52:01.781567281 +0000 UTC m=+913.194883328" lastFinishedPulling="2026-03-17 15:52:02.859394473 +0000 UTC m=+914.272710520" observedRunningTime="2026-03-17 15:52:03.489830329 +0000 UTC m=+914.903146396" watchObservedRunningTime="2026-03-17 15:52:03.516378992 +0000 UTC m=+914.929695039" Mar 17 15:52:04 crc kubenswrapper[4767]: I0317 15:52:04.475994 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4ed68ee-0956-4d37-9d3b-f1904ff19451" containerID="63135d6512e4bb5413b9841549fb5ec1a52f9e8411d5691cb6707e9465c12312" exitCode=0 Mar 17 15:52:04 crc kubenswrapper[4767]: I0317 15:52:04.476100 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562712-r589g" event={"ID":"f4ed68ee-0956-4d37-9d3b-f1904ff19451","Type":"ContainerDied","Data":"63135d6512e4bb5413b9841549fb5ec1a52f9e8411d5691cb6707e9465c12312"} Mar 17 15:52:04 crc kubenswrapper[4767]: I0317 15:52:04.494986 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-msl8c" podStartSLOduration=4.486027766 podStartE2EDuration="10.494961294s" podCreationTimestamp="2026-03-17 15:51:54 +0000 UTC" firstStartedPulling="2026-03-17 15:51:56.885916207 +0000 UTC m=+908.299232254" lastFinishedPulling="2026-03-17 15:52:02.894849735 +0000 UTC m=+914.308165782" observedRunningTime="2026-03-17 15:52:03.519004458 +0000 UTC m=+914.932320525" watchObservedRunningTime="2026-03-17 15:52:04.494961294 +0000 UTC m=+915.908277331" Mar 17 15:52:05 crc kubenswrapper[4767]: I0317 15:52:05.550124 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:52:05 crc kubenswrapper[4767]: I0317 15:52:05.550604 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:52:06 crc kubenswrapper[4767]: I0317 15:52:06.345875 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562712-r589g" Mar 17 15:52:06 crc kubenswrapper[4767]: I0317 15:52:06.455561 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4hbn\" (UniqueName: \"kubernetes.io/projected/f4ed68ee-0956-4d37-9d3b-f1904ff19451-kube-api-access-k4hbn\") pod \"f4ed68ee-0956-4d37-9d3b-f1904ff19451\" (UID: \"f4ed68ee-0956-4d37-9d3b-f1904ff19451\") " Mar 17 15:52:06 crc kubenswrapper[4767]: I0317 15:52:06.475512 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4ed68ee-0956-4d37-9d3b-f1904ff19451-kube-api-access-k4hbn" (OuterVolumeSpecName: "kube-api-access-k4hbn") pod "f4ed68ee-0956-4d37-9d3b-f1904ff19451" (UID: "f4ed68ee-0956-4d37-9d3b-f1904ff19451"). InnerVolumeSpecName "kube-api-access-k4hbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:52:06 crc kubenswrapper[4767]: I0317 15:52:06.496326 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562712-r589g" event={"ID":"f4ed68ee-0956-4d37-9d3b-f1904ff19451","Type":"ContainerDied","Data":"5da8c9466851405b6631cb9b499584cce3f1f801525aee4864ee89ac6214196d"} Mar 17 15:52:06 crc kubenswrapper[4767]: I0317 15:52:06.496385 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5da8c9466851405b6631cb9b499584cce3f1f801525aee4864ee89ac6214196d" Mar 17 15:52:06 crc kubenswrapper[4767]: I0317 15:52:06.496424 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562712-r589g" Mar 17 15:52:06 crc kubenswrapper[4767]: I0317 15:52:06.555393 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562706-jzpkc"] Mar 17 15:52:06 crc kubenswrapper[4767]: I0317 15:52:06.558060 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4hbn\" (UniqueName: \"kubernetes.io/projected/f4ed68ee-0956-4d37-9d3b-f1904ff19451-kube-api-access-k4hbn\") on node \"crc\" DevicePath \"\"" Mar 17 15:52:06 crc kubenswrapper[4767]: I0317 15:52:06.560934 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562706-jzpkc"] Mar 17 15:52:06 crc kubenswrapper[4767]: I0317 15:52:06.770277 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-msl8c" podUID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerName="registry-server" probeResult="failure" output=< Mar 17 15:52:06 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 15:52:06 crc kubenswrapper[4767]: > Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.160348 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-66689c4bbf-gv6qd"] Mar 17 15:52:07 crc kubenswrapper[4767]: E0317 15:52:07.160747 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4ed68ee-0956-4d37-9d3b-f1904ff19451" containerName="oc" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.160773 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4ed68ee-0956-4d37-9d3b-f1904ff19451" containerName="oc" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.160998 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4ed68ee-0956-4d37-9d3b-f1904ff19451" containerName="oc" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.161763 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-66689c4bbf-gv6qd" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.166834 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.166888 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.166850 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-nb48k" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.174830 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-66689c4bbf-gv6qd"] Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.272346 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sgzw\" (UniqueName: \"kubernetes.io/projected/b117c227-2f96-4562-99f1-eb7c1a080059-kube-api-access-2sgzw\") pod \"cluster-logging-operator-66689c4bbf-gv6qd\" (UID: \"b117c227-2f96-4562-99f1-eb7c1a080059\") " pod="openshift-logging/cluster-logging-operator-66689c4bbf-gv6qd" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.364073 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21009afb-c8b8-4301-a400-8e193d50c6a9" path="/var/lib/kubelet/pods/21009afb-c8b8-4301-a400-8e193d50c6a9/volumes" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.374224 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sgzw\" (UniqueName: \"kubernetes.io/projected/b117c227-2f96-4562-99f1-eb7c1a080059-kube-api-access-2sgzw\") pod \"cluster-logging-operator-66689c4bbf-gv6qd\" (UID: \"b117c227-2f96-4562-99f1-eb7c1a080059\") " pod="openshift-logging/cluster-logging-operator-66689c4bbf-gv6qd" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.400078 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sgzw\" (UniqueName: \"kubernetes.io/projected/b117c227-2f96-4562-99f1-eb7c1a080059-kube-api-access-2sgzw\") pod \"cluster-logging-operator-66689c4bbf-gv6qd\" (UID: \"b117c227-2f96-4562-99f1-eb7c1a080059\") " pod="openshift-logging/cluster-logging-operator-66689c4bbf-gv6qd" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.504033 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-66689c4bbf-gv6qd" Mar 17 15:52:07 crc kubenswrapper[4767]: I0317 15:52:07.768180 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-66689c4bbf-gv6qd"] Mar 17 15:52:08 crc kubenswrapper[4767]: I0317 15:52:08.526459 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-66689c4bbf-gv6qd" event={"ID":"b117c227-2f96-4562-99f1-eb7c1a080059","Type":"ContainerStarted","Data":"720ec4dafc1172bbeec81393a3aa77c15601710035c7b611f9d7274e00f35cfa"} Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.102107 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm"] Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.110639 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.114701 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.114735 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-g28mp" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.114724 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.116070 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.117065 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.118251 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.138027 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm"] Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.214514 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/6f682203-ee25-4b31-957b-26d8148feda0-manager-config\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.214571 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpxf9\" (UniqueName: \"kubernetes.io/projected/6f682203-ee25-4b31-957b-26d8148feda0-kube-api-access-lpxf9\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.214596 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6f682203-ee25-4b31-957b-26d8148feda0-apiservice-cert\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.214645 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6f682203-ee25-4b31-957b-26d8148feda0-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.214687 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6f682203-ee25-4b31-957b-26d8148feda0-webhook-cert\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.316501 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6f682203-ee25-4b31-957b-26d8148feda0-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.316620 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6f682203-ee25-4b31-957b-26d8148feda0-webhook-cert\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.316737 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/6f682203-ee25-4b31-957b-26d8148feda0-manager-config\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.316763 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpxf9\" (UniqueName: \"kubernetes.io/projected/6f682203-ee25-4b31-957b-26d8148feda0-kube-api-access-lpxf9\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.316787 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6f682203-ee25-4b31-957b-26d8148feda0-apiservice-cert\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.318966 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/6f682203-ee25-4b31-957b-26d8148feda0-manager-config\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.324904 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6f682203-ee25-4b31-957b-26d8148feda0-webhook-cert\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.324948 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6f682203-ee25-4b31-957b-26d8148feda0-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.328057 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6f682203-ee25-4b31-957b-26d8148feda0-apiservice-cert\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.345694 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpxf9\" (UniqueName: \"kubernetes.io/projected/6f682203-ee25-4b31-957b-26d8148feda0-kube-api-access-lpxf9\") pod \"loki-operator-controller-manager-5dcc96cb6-rv7xm\" (UID: \"6f682203-ee25-4b31-957b-26d8148feda0\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:13 crc kubenswrapper[4767]: I0317 15:52:13.431811 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:15 crc kubenswrapper[4767]: I0317 15:52:15.599601 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:52:15 crc kubenswrapper[4767]: I0317 15:52:15.671221 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:52:18 crc kubenswrapper[4767]: I0317 15:52:18.934061 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-msl8c"] Mar 17 15:52:18 crc kubenswrapper[4767]: I0317 15:52:18.934974 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-msl8c" podUID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerName="registry-server" containerID="cri-o://0b5d543001dab9e12a0aa329c64fbc38b3a1ef67afebfbe7607ec1f78683dd54" gracePeriod=2 Mar 17 15:52:19 crc kubenswrapper[4767]: I0317 15:52:19.789649 4767 generic.go:334] "Generic (PLEG): container finished" podID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerID="0b5d543001dab9e12a0aa329c64fbc38b3a1ef67afebfbe7607ec1f78683dd54" exitCode=0 Mar 17 15:52:19 crc kubenswrapper[4767]: I0317 15:52:19.790510 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msl8c" event={"ID":"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b","Type":"ContainerDied","Data":"0b5d543001dab9e12a0aa329c64fbc38b3a1ef67afebfbe7607ec1f78683dd54"} Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.339252 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.370842 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm"] Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.389253 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-utilities\") pod \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.389325 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57htg\" (UniqueName: \"kubernetes.io/projected/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-kube-api-access-57htg\") pod \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.389382 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-catalog-content\") pod \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\" (UID: \"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b\") " Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.391164 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-utilities" (OuterVolumeSpecName: "utilities") pod "7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" (UID: "7ae678ad-31d9-4eb5-84ba-ab0fbd24313b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.402519 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-kube-api-access-57htg" (OuterVolumeSpecName: "kube-api-access-57htg") pod "7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" (UID: "7ae678ad-31d9-4eb5-84ba-ab0fbd24313b"). InnerVolumeSpecName "kube-api-access-57htg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.490913 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.490998 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57htg\" (UniqueName: \"kubernetes.io/projected/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-kube-api-access-57htg\") on node \"crc\" DevicePath \"\"" Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.569950 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" (UID: "7ae678ad-31d9-4eb5-84ba-ab0fbd24313b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.592117 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.816435 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" event={"ID":"6f682203-ee25-4b31-957b-26d8148feda0","Type":"ContainerStarted","Data":"a106426411f28bc4cc5ed3bf0919dbff1393be8b53127fb1079b2e22103d77e1"} Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.819672 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-66689c4bbf-gv6qd" event={"ID":"b117c227-2f96-4562-99f1-eb7c1a080059","Type":"ContainerStarted","Data":"386e90695434245699c51de296f4bcbc7e14c7d2a9ede932f8ccc364c1bceec3"} Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.827661 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msl8c" event={"ID":"7ae678ad-31d9-4eb5-84ba-ab0fbd24313b","Type":"ContainerDied","Data":"100e6816cb0f33163699474491989651207e22422601307fc02ce93f66fd980e"} Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.827736 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-msl8c" Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.827774 4767 scope.go:117] "RemoveContainer" containerID="0b5d543001dab9e12a0aa329c64fbc38b3a1ef67afebfbe7607ec1f78683dd54" Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.855679 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-66689c4bbf-gv6qd" podStartSLOduration=1.953968342 podStartE2EDuration="13.855644091s" podCreationTimestamp="2026-03-17 15:52:07 +0000 UTC" firstStartedPulling="2026-03-17 15:52:07.791214798 +0000 UTC m=+919.204530845" lastFinishedPulling="2026-03-17 15:52:19.692890547 +0000 UTC m=+931.106206594" observedRunningTime="2026-03-17 15:52:20.849063069 +0000 UTC m=+932.262379116" watchObservedRunningTime="2026-03-17 15:52:20.855644091 +0000 UTC m=+932.268960138" Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.862241 4767 scope.go:117] "RemoveContainer" containerID="eb96fe5161d6c9c53ec50e1e3626bbf6d862d764960a33730a3571ef82cbeef2" Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.885062 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-msl8c"] Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.893258 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-msl8c"] Mar 17 15:52:20 crc kubenswrapper[4767]: I0317 15:52:20.897409 4767 scope.go:117] "RemoveContainer" containerID="a21f8cec437b64ce1e106119a6a73e87fc9653ca5cf8833a0907b7b9f1635c4d" Mar 17 15:52:21 crc kubenswrapper[4767]: I0317 15:52:21.364221 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" path="/var/lib/kubelet/pods/7ae678ad-31d9-4eb5-84ba-ab0fbd24313b/volumes" Mar 17 15:52:26 crc kubenswrapper[4767]: I0317 15:52:26.893847 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" event={"ID":"6f682203-ee25-4b31-957b-26d8148feda0","Type":"ContainerStarted","Data":"e92b976b8865fc5db41829a9fa6f48341ac42c83d5319bca392cb1fe37e46e93"} Mar 17 15:52:30 crc kubenswrapper[4767]: I0317 15:52:30.814526 4767 scope.go:117] "RemoveContainer" containerID="d225fc6960a166ccd176dfa28e583ae0a4c957489c67ff5dc3a46a1a412444c9" Mar 17 15:52:33 crc kubenswrapper[4767]: I0317 15:52:33.969382 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" event={"ID":"6f682203-ee25-4b31-957b-26d8148feda0","Type":"ContainerStarted","Data":"32075afe5bb1ca1378395490e025e1577b940f3123aeca80c1d83a8ce763ad0a"} Mar 17 15:52:33 crc kubenswrapper[4767]: I0317 15:52:33.972258 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:33 crc kubenswrapper[4767]: I0317 15:52:33.979473 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 15:52:34 crc kubenswrapper[4767]: I0317 15:52:34.001146 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" podStartSLOduration=8.544730944 podStartE2EDuration="21.001123043s" podCreationTimestamp="2026-03-17 15:52:13 +0000 UTC" firstStartedPulling="2026-03-17 15:52:20.373832652 +0000 UTC m=+931.787148699" lastFinishedPulling="2026-03-17 15:52:32.830224751 +0000 UTC m=+944.243540798" observedRunningTime="2026-03-17 15:52:34.000495645 +0000 UTC m=+945.413811702" watchObservedRunningTime="2026-03-17 15:52:34.001123043 +0000 UTC m=+945.414439110" Mar 17 15:52:38 crc kubenswrapper[4767]: I0317 15:52:38.766513 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Mar 17 15:52:38 crc kubenswrapper[4767]: E0317 15:52:38.771901 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerName="extract-content" Mar 17 15:52:38 crc kubenswrapper[4767]: I0317 15:52:38.771973 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerName="extract-content" Mar 17 15:52:38 crc kubenswrapper[4767]: E0317 15:52:38.772005 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerName="registry-server" Mar 17 15:52:38 crc kubenswrapper[4767]: I0317 15:52:38.772014 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerName="registry-server" Mar 17 15:52:38 crc kubenswrapper[4767]: E0317 15:52:38.772028 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerName="extract-utilities" Mar 17 15:52:38 crc kubenswrapper[4767]: I0317 15:52:38.772036 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerName="extract-utilities" Mar 17 15:52:38 crc kubenswrapper[4767]: I0317 15:52:38.772227 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ae678ad-31d9-4eb5-84ba-ab0fbd24313b" containerName="registry-server" Mar 17 15:52:38 crc kubenswrapper[4767]: I0317 15:52:38.773066 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Mar 17 15:52:38 crc kubenswrapper[4767]: I0317 15:52:38.788160 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Mar 17 15:52:38 crc kubenswrapper[4767]: I0317 15:52:38.788445 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Mar 17 15:52:38 crc kubenswrapper[4767]: I0317 15:52:38.783305 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Mar 17 15:52:38 crc kubenswrapper[4767]: I0317 15:52:38.947299 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnzbf\" (UniqueName: \"kubernetes.io/projected/c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5-kube-api-access-lnzbf\") pod \"minio\" (UID: \"c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5\") " pod="minio-dev/minio" Mar 17 15:52:38 crc kubenswrapper[4767]: I0317 15:52:38.947449 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-48777082-ec86-4701-a8c1-bf2814c258c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48777082-ec86-4701-a8c1-bf2814c258c1\") pod \"minio\" (UID: \"c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5\") " pod="minio-dev/minio" Mar 17 15:52:39 crc kubenswrapper[4767]: I0317 15:52:39.048614 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnzbf\" (UniqueName: \"kubernetes.io/projected/c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5-kube-api-access-lnzbf\") pod \"minio\" (UID: \"c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5\") " pod="minio-dev/minio" Mar 17 15:52:39 crc kubenswrapper[4767]: I0317 15:52:39.048748 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-48777082-ec86-4701-a8c1-bf2814c258c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48777082-ec86-4701-a8c1-bf2814c258c1\") pod \"minio\" (UID: \"c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5\") " pod="minio-dev/minio" Mar 17 15:52:39 crc kubenswrapper[4767]: I0317 15:52:39.052639 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:52:39 crc kubenswrapper[4767]: I0317 15:52:39.052672 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-48777082-ec86-4701-a8c1-bf2814c258c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48777082-ec86-4701-a8c1-bf2814c258c1\") pod \"minio\" (UID: \"c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/07a482b46a9bfde292c5c35f47c459c089e28f92d6c563cae57a73bfd8f105f2/globalmount\"" pod="minio-dev/minio" Mar 17 15:52:39 crc kubenswrapper[4767]: I0317 15:52:39.078648 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnzbf\" (UniqueName: \"kubernetes.io/projected/c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5-kube-api-access-lnzbf\") pod \"minio\" (UID: \"c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5\") " pod="minio-dev/minio" Mar 17 15:52:39 crc kubenswrapper[4767]: I0317 15:52:39.095938 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-48777082-ec86-4701-a8c1-bf2814c258c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48777082-ec86-4701-a8c1-bf2814c258c1\") pod \"minio\" (UID: \"c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5\") " pod="minio-dev/minio" Mar 17 15:52:39 crc kubenswrapper[4767]: I0317 15:52:39.124194 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Mar 17 15:52:39 crc kubenswrapper[4767]: I0317 15:52:39.653582 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Mar 17 15:52:40 crc kubenswrapper[4767]: I0317 15:52:40.015152 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5","Type":"ContainerStarted","Data":"e4d11af4b2144f4a6ee371b7d6a0559eb829bdd3a2234cbd5893495f5c12bb7a"} Mar 17 15:52:46 crc kubenswrapper[4767]: I0317 15:52:46.341327 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"c6dfbfcd-6ebb-45e0-9840-4f0a6aeb13b5","Type":"ContainerStarted","Data":"954eaef596d67153809f5fda6a05c37ad300c05b9e6225fedd1dfc962737222c"} Mar 17 15:52:46 crc kubenswrapper[4767]: I0317 15:52:46.363368 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.741600028 podStartE2EDuration="10.363334551s" podCreationTimestamp="2026-03-17 15:52:36 +0000 UTC" firstStartedPulling="2026-03-17 15:52:39.670763109 +0000 UTC m=+951.084079156" lastFinishedPulling="2026-03-17 15:52:45.292497632 +0000 UTC m=+956.705813679" observedRunningTime="2026-03-17 15:52:46.358813459 +0000 UTC m=+957.772129526" watchObservedRunningTime="2026-03-17 15:52:46.363334551 +0000 UTC m=+957.776650608" Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.620235 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v75gc"] Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.623371 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.634530 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v75gc"] Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.810651 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd4c2\" (UniqueName: \"kubernetes.io/projected/04d79068-7ef7-42f8-81e5-c5211ec20c12-kube-api-access-gd4c2\") pod \"community-operators-v75gc\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.810793 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-catalog-content\") pod \"community-operators-v75gc\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.811278 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-utilities\") pod \"community-operators-v75gc\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.912530 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-catalog-content\") pod \"community-operators-v75gc\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.912664 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-utilities\") pod \"community-operators-v75gc\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.912704 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd4c2\" (UniqueName: \"kubernetes.io/projected/04d79068-7ef7-42f8-81e5-c5211ec20c12-kube-api-access-gd4c2\") pod \"community-operators-v75gc\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.913303 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-utilities\") pod \"community-operators-v75gc\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.913582 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-catalog-content\") pod \"community-operators-v75gc\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.946939 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd4c2\" (UniqueName: \"kubernetes.io/projected/04d79068-7ef7-42f8-81e5-c5211ec20c12-kube-api-access-gd4c2\") pod \"community-operators-v75gc\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:49 crc kubenswrapper[4767]: I0317 15:52:49.955491 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:50 crc kubenswrapper[4767]: I0317 15:52:50.840691 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v75gc"] Mar 17 15:52:50 crc kubenswrapper[4767]: W0317 15:52:50.855612 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04d79068_7ef7_42f8_81e5_c5211ec20c12.slice/crio-2e9cf3eb3d4a8ca2977bfdf7602b67857ddcad2fad149b2567630442a2559864 WatchSource:0}: Error finding container 2e9cf3eb3d4a8ca2977bfdf7602b67857ddcad2fad149b2567630442a2559864: Status 404 returned error can't find the container with id 2e9cf3eb3d4a8ca2977bfdf7602b67857ddcad2fad149b2567630442a2559864 Mar 17 15:52:51 crc kubenswrapper[4767]: I0317 15:52:51.383997 4767 generic.go:334] "Generic (PLEG): container finished" podID="04d79068-7ef7-42f8-81e5-c5211ec20c12" containerID="84046a4dbc6d48fb9bbf0fc1dd851ebd8844dbef28561f13c67cce6473db8870" exitCode=0 Mar 17 15:52:51 crc kubenswrapper[4767]: I0317 15:52:51.384060 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v75gc" event={"ID":"04d79068-7ef7-42f8-81e5-c5211ec20c12","Type":"ContainerDied","Data":"84046a4dbc6d48fb9bbf0fc1dd851ebd8844dbef28561f13c67cce6473db8870"} Mar 17 15:52:51 crc kubenswrapper[4767]: I0317 15:52:51.384151 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v75gc" event={"ID":"04d79068-7ef7-42f8-81e5-c5211ec20c12","Type":"ContainerStarted","Data":"2e9cf3eb3d4a8ca2977bfdf7602b67857ddcad2fad149b2567630442a2559864"} Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:52.988348 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr"] Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:52.990111 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:52.992562 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:52.993012 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:52.993055 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-hp4fk" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:52.993350 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:52.994706 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.099139 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/28ac74bb-3f68-478d-8f50-acf61b2e8223-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.099241 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d49gp\" (UniqueName: \"kubernetes.io/projected/28ac74bb-3f68-478d-8f50-acf61b2e8223-kube-api-access-d49gp\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.099290 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28ac74bb-3f68-478d-8f50-acf61b2e8223-config\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.099468 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/28ac74bb-3f68-478d-8f50-acf61b2e8223-logging-loki-distributor-http\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.099508 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28ac74bb-3f68-478d-8f50-acf61b2e8223-logging-loki-ca-bundle\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.129927 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr"] Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.200922 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/28ac74bb-3f68-478d-8f50-acf61b2e8223-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.201637 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d49gp\" (UniqueName: \"kubernetes.io/projected/28ac74bb-3f68-478d-8f50-acf61b2e8223-kube-api-access-d49gp\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.201691 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28ac74bb-3f68-478d-8f50-acf61b2e8223-config\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.201780 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/28ac74bb-3f68-478d-8f50-acf61b2e8223-logging-loki-distributor-http\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.201820 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28ac74bb-3f68-478d-8f50-acf61b2e8223-logging-loki-ca-bundle\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.210229 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28ac74bb-3f68-478d-8f50-acf61b2e8223-logging-loki-ca-bundle\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.213665 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28ac74bb-3f68-478d-8f50-acf61b2e8223-config\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.225506 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd"] Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.226522 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.231401 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/28ac74bb-3f68-478d-8f50-acf61b2e8223-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.233149 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/28ac74bb-3f68-478d-8f50-acf61b2e8223-logging-loki-distributor-http\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.233390 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.233592 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.241867 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.246766 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd"] Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.276817 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d49gp\" (UniqueName: \"kubernetes.io/projected/28ac74bb-3f68-478d-8f50-acf61b2e8223-kube-api-access-d49gp\") pod \"logging-loki-distributor-9c6b6d984-q9dnr\" (UID: \"28ac74bb-3f68-478d-8f50-acf61b2e8223\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.306676 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-ca-bundle\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.306778 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5p6b\" (UniqueName: \"kubernetes.io/projected/66451b46-d0f5-4037-aa4d-abf0143b2ef8-kube-api-access-s5p6b\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.306805 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-s3\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.306849 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66451b46-d0f5-4037-aa4d-abf0143b2ef8-config\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.306967 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-querier-grpc\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.307017 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-querier-http\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.487720 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.488099 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-ca-bundle\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.488262 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5p6b\" (UniqueName: \"kubernetes.io/projected/66451b46-d0f5-4037-aa4d-abf0143b2ef8-kube-api-access-s5p6b\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.488295 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-s3\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.488363 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66451b46-d0f5-4037-aa4d-abf0143b2ef8-config\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.488511 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-querier-grpc\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.488578 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-querier-http\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.494106 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66451b46-d0f5-4037-aa4d-abf0143b2ef8-config\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.502159 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-querier-http\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.503552 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-s3\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.517781 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-ca-bundle\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.523435 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/66451b46-d0f5-4037-aa4d-abf0143b2ef8-logging-loki-querier-grpc\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.568990 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5p6b\" (UniqueName: \"kubernetes.io/projected/66451b46-d0f5-4037-aa4d-abf0143b2ef8-kube-api-access-s5p6b\") pod \"logging-loki-querier-6dcbdf8bb8-679nd\" (UID: \"66451b46-d0f5-4037-aa4d-abf0143b2ef8\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.584630 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v75gc" event={"ID":"04d79068-7ef7-42f8-81e5-c5211ec20c12","Type":"ContainerStarted","Data":"4c91a5e32f216332ae10702575de4cfcfe0ed53ce37a9c1c52193afe8298dd32"} Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.631687 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt"] Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.654244 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.659584 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt"] Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.659749 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.680529 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.680815 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.718794 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.718932 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.718969 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.719004 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvb8k\" (UniqueName: \"kubernetes.io/projected/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-kube-api-access-xvb8k\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.719088 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-config\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.821421 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.821528 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvb8k\" (UniqueName: \"kubernetes.io/projected/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-kube-api-access-xvb8k\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.821615 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-config\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.821670 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.821801 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.825543 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.825751 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-config\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.832786 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:53 crc kubenswrapper[4767]: I0317 15:52:53.855596 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.162821 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvb8k\" (UniqueName: \"kubernetes.io/projected/ce1ebb9b-1558-4190-9e47-4ef6ba3160da-kube-api-access-xvb8k\") pod \"logging-loki-query-frontend-ff66c4dc9-4jtkt\" (UID: \"ce1ebb9b-1558-4190-9e47-4ef6ba3160da\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.170439 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-76dddc4d57-7gw68"] Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.172047 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.343764 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-lw4qq" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.344014 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.344120 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.344487 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.344631 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.345006 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.374206 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4"] Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.376090 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.385033 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.392320 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-76dddc4d57-7gw68"] Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.412459 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.413913 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.419793 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.420914 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.424554 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.425871 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.428173 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4"] Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.428741 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.429539 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.446752 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448269 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448319 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448354 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhdhx\" (UniqueName: \"kubernetes.io/projected/560edfa4-8a88-4c9b-8b31-e61f93050c15-kube-api-access-fhdhx\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448380 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448395 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448424 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-tenants\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448447 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-lokistack-gateway\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448462 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-rbac\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448484 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-tenants\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448499 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448649 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-tls-secret\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448679 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-rbac\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448716 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448738 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffgvp\" (UniqueName: \"kubernetes.io/projected/9c622cae-f125-4675-9bb5-d2dd999edfa3-kube-api-access-ffgvp\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448760 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-tls-secret\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.448780 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-lokistack-gateway\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.450834 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.551777 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552619 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552660 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552685 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f972f374-ac6a-4b03-93cd-7b6972737681\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f972f374-ac6a-4b03-93cd-7b6972737681\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552716 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552747 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552789 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lldt9\" (UniqueName: \"kubernetes.io/projected/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-kube-api-access-lldt9\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552825 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-rbac\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552849 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-tls-secret\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552878 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552908 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffgvp\" (UniqueName: \"kubernetes.io/projected/9c622cae-f125-4675-9bb5-d2dd999edfa3-kube-api-access-ffgvp\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552936 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-tls-secret\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552960 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-lokistack-gateway\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.552983 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2389106-04cc-4871-b544-f8a677a6405a-config\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.553012 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c266a566-feca-4e5b-98f3-ad0dcc359290\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c266a566-feca-4e5b-98f3-ad0dcc359290\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.553095 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.553143 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxtgj\" (UniqueName: \"kubernetes.io/projected/a2389106-04cc-4871-b544-f8a677a6405a-kube-api-access-wxtgj\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: E0317 15:52:54.553194 4767 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Mar 17 15:52:54 crc kubenswrapper[4767]: E0317 15:52:54.553321 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-tls-secret podName:560edfa4-8a88-4c9b-8b31-e61f93050c15 nodeName:}" failed. No retries permitted until 2026-03-17 15:52:55.053299809 +0000 UTC m=+966.466615856 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-tls-secret") pod "logging-loki-gateway-76dddc4d57-8q5h4" (UID: "560edfa4-8a88-4c9b-8b31-e61f93050c15") : secret "logging-loki-gateway-http" not found Mar 17 15:52:54 crc kubenswrapper[4767]: E0317 15:52:54.553577 4767 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Mar 17 15:52:54 crc kubenswrapper[4767]: E0317 15:52:54.553640 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-tls-secret podName:9c622cae-f125-4675-9bb5-d2dd999edfa3 nodeName:}" failed. No retries permitted until 2026-03-17 15:52:55.053615618 +0000 UTC m=+966.466931665 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-tls-secret") pod "logging-loki-gateway-76dddc4d57-7gw68" (UID: "9c622cae-f125-4675-9bb5-d2dd999edfa3") : secret "logging-loki-gateway-http" not found Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.553227 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.555254 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-lokistack-gateway\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.555717 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhdhx\" (UniqueName: \"kubernetes.io/projected/560edfa4-8a88-4c9b-8b31-e61f93050c15-kube-api-access-fhdhx\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.555842 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.555899 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.555981 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9d48a2f0-daf9-4db9-9433-da5fea38922f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9d48a2f0-daf9-4db9-9433-da5fea38922f\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.556031 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.556089 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.556134 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-tenants\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.556185 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.556349 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-lokistack-gateway\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.556383 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-rbac\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.556420 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.556423 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-tenants\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.556500 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.556555 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-config\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.559097 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.560073 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/560edfa4-8a88-4c9b-8b31-e61f93050c15-rbac\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.568488 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-logging-loki-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.570093 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-tenants\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.572213 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-rbac\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.576865 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.577029 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffgvp\" (UniqueName: \"kubernetes.io/projected/9c622cae-f125-4675-9bb5-d2dd999edfa3-kube-api-access-ffgvp\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.577332 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.584127 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhdhx\" (UniqueName: \"kubernetes.io/projected/560edfa4-8a88-4c9b-8b31-e61f93050c15-kube-api-access-fhdhx\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.607166 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.611026 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/9c622cae-f125-4675-9bb5-d2dd999edfa3-lokistack-gateway\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.617991 4767 generic.go:334] "Generic (PLEG): container finished" podID="04d79068-7ef7-42f8-81e5-c5211ec20c12" containerID="4c91a5e32f216332ae10702575de4cfcfe0ed53ce37a9c1c52193afe8298dd32" exitCode=0 Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.618071 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v75gc" event={"ID":"04d79068-7ef7-42f8-81e5-c5211ec20c12","Type":"ContainerDied","Data":"4c91a5e32f216332ae10702575de4cfcfe0ed53ce37a9c1c52193afe8298dd32"} Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.618579 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-tenants\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.658532 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxtgj\" (UniqueName: \"kubernetes.io/projected/a2389106-04cc-4871-b544-f8a677a6405a-kube-api-access-wxtgj\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.658727 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9d48a2f0-daf9-4db9-9433-da5fea38922f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9d48a2f0-daf9-4db9-9433-da5fea38922f\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.658755 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.658784 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.658816 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.658892 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-config\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.658918 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.658957 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.658977 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.659000 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f972f374-ac6a-4b03-93cd-7b6972737681\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f972f374-ac6a-4b03-93cd-7b6972737681\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.659020 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.659053 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.659088 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lldt9\" (UniqueName: \"kubernetes.io/projected/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-kube-api-access-lldt9\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.659192 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2389106-04cc-4871-b544-f8a677a6405a-config\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.659243 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c266a566-feca-4e5b-98f3-ad0dcc359290\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c266a566-feca-4e5b-98f3-ad0dcc359290\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.663543 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.665740 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-config\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.666142 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2389106-04cc-4871-b544-f8a677a6405a-config\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.669112 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.669184 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c266a566-feca-4e5b-98f3-ad0dcc359290\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c266a566-feca-4e5b-98f3-ad0dcc359290\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1b3143a46f7de819fc8b8dbdc7c73d50ff16eff0185f84864e09afb6b52f60b1/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.669601 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.669655 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9d48a2f0-daf9-4db9-9433-da5fea38922f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9d48a2f0-daf9-4db9-9433-da5fea38922f\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/30077ebe717424edb25188989d5f1d6567c4117dde0bb59a3f83d33af3078fb0/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.670448 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.671456 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.671484 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f972f374-ac6a-4b03-93cd-7b6972737681\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f972f374-ac6a-4b03-93cd-7b6972737681\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/85e2ab10157bfe3ee3b3cde3829a7d77ec46e5fe660935be018e325b3f73d40b/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.673728 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.678409 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.682071 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.682654 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/a2389106-04cc-4871-b544-f8a677a6405a-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.685102 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.687195 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.689128 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lldt9\" (UniqueName: \"kubernetes.io/projected/27c143b3-0f5b-4265-a2e2-d4302a2f3c70-kube-api-access-lldt9\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.695259 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxtgj\" (UniqueName: \"kubernetes.io/projected/a2389106-04cc-4871-b544-f8a677a6405a-kube-api-access-wxtgj\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.737134 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.738313 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.742741 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.742925 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.760896 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.761312 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.761451 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.761564 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.761667 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89ce8746-1721-4dec-9742-0abc2f498de3-config\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.761775 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b00d33e6-8779-4c5f-ad6b-1c70e9f2cbb1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b00d33e6-8779-4c5f-ad6b-1c70e9f2cbb1\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.761872 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55298\" (UniqueName: \"kubernetes.io/projected/89ce8746-1721-4dec-9742-0abc2f498de3-kube-api-access-55298\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.796820 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f972f374-ac6a-4b03-93cd-7b6972737681\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f972f374-ac6a-4b03-93cd-7b6972737681\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.820292 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.854160 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c266a566-feca-4e5b-98f3-ad0dcc359290\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c266a566-feca-4e5b-98f3-ad0dcc359290\") pod \"logging-loki-ingester-0\" (UID: \"a2389106-04cc-4871-b544-f8a677a6405a\") " pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.861457 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9d48a2f0-daf9-4db9-9433-da5fea38922f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9d48a2f0-daf9-4db9-9433-da5fea38922f\") pod \"logging-loki-compactor-0\" (UID: \"27c143b3-0f5b-4265-a2e2-d4302a2f3c70\") " pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.875411 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.875505 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.875560 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.875607 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89ce8746-1721-4dec-9742-0abc2f498de3-config\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.875642 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b00d33e6-8779-4c5f-ad6b-1c70e9f2cbb1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b00d33e6-8779-4c5f-ad6b-1c70e9f2cbb1\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.875723 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55298\" (UniqueName: \"kubernetes.io/projected/89ce8746-1721-4dec-9742-0abc2f498de3-kube-api-access-55298\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.876018 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.884924 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89ce8746-1721-4dec-9742-0abc2f498de3-config\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.887932 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.898805 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.900807 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.914486 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.914567 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b00d33e6-8779-4c5f-ad6b-1c70e9f2cbb1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b00d33e6-8779-4c5f-ad6b-1c70e9f2cbb1\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c537d4971f86e70f1dc443629ddc0d379fb9d813d38b9061e5d920eecd884a94/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.925220 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/89ce8746-1721-4dec-9742-0abc2f498de3-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.941929 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55298\" (UniqueName: \"kubernetes.io/projected/89ce8746-1721-4dec-9742-0abc2f498de3-kube-api-access-55298\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.945286 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr"] Mar 17 15:52:54 crc kubenswrapper[4767]: I0317 15:52:54.964491 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd"] Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.023281 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b00d33e6-8779-4c5f-ad6b-1c70e9f2cbb1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b00d33e6-8779-4c5f-ad6b-1c70e9f2cbb1\") pod \"logging-loki-index-gateway-0\" (UID: \"89ce8746-1721-4dec-9742-0abc2f498de3\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.079955 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.089690 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-tls-secret\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.090327 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-tls-secret\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.107478 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/560edfa4-8a88-4c9b-8b31-e61f93050c15-tls-secret\") pod \"logging-loki-gateway-76dddc4d57-8q5h4\" (UID: \"560edfa4-8a88-4c9b-8b31-e61f93050c15\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.107741 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/9c622cae-f125-4675-9bb5-d2dd999edfa3-tls-secret\") pod \"logging-loki-gateway-76dddc4d57-7gw68\" (UID: \"9c622cae-f125-4675-9bb5-d2dd999edfa3\") " pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.121871 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.135572 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.153954 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.165007 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.345283 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt"] Mar 17 15:52:55 crc kubenswrapper[4767]: W0317 15:52:55.444628 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce1ebb9b_1558_4190_9e47_4ef6ba3160da.slice/crio-9fe46788fc06a24cc4835a332cfcf4df8a6e97f6e36518be729b704d60652d9a WatchSource:0}: Error finding container 9fe46788fc06a24cc4835a332cfcf4df8a6e97f6e36518be729b704d60652d9a: Status 404 returned error can't find the container with id 9fe46788fc06a24cc4835a332cfcf4df8a6e97f6e36518be729b704d60652d9a Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.526795 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.649368 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" event={"ID":"ce1ebb9b-1558-4190-9e47-4ef6ba3160da","Type":"ContainerStarted","Data":"9fe46788fc06a24cc4835a332cfcf4df8a6e97f6e36518be729b704d60652d9a"} Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.656447 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" event={"ID":"28ac74bb-3f68-478d-8f50-acf61b2e8223","Type":"ContainerStarted","Data":"adcf5114c2b0ecc762d7442b4e2b223b082c59f4403d90424848deca5a09bedb"} Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.657448 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" event={"ID":"66451b46-d0f5-4037-aa4d-abf0143b2ef8","Type":"ContainerStarted","Data":"c1d5d01bb0e55ee1039684dac6fc809bfba446aafe7a633e0a5c67f48e1c78ff"} Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.659701 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"89ce8746-1721-4dec-9742-0abc2f498de3","Type":"ContainerStarted","Data":"18cab8fbcdf72ce7bcb8c013314ceef4055948abe68b195c470068cc4a0c0913"} Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.676083 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v75gc" podStartSLOduration=2.737718231 podStartE2EDuration="6.676044608s" podCreationTimestamp="2026-03-17 15:52:49 +0000 UTC" firstStartedPulling="2026-03-17 15:52:51.387419662 +0000 UTC m=+962.800735719" lastFinishedPulling="2026-03-17 15:52:55.325746049 +0000 UTC m=+966.739062096" observedRunningTime="2026-03-17 15:52:55.675317547 +0000 UTC m=+967.088633614" watchObservedRunningTime="2026-03-17 15:52:55.676044608 +0000 UTC m=+967.089360655" Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.791461 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-76dddc4d57-7gw68"] Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.900240 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4"] Mar 17 15:52:55 crc kubenswrapper[4767]: I0317 15:52:55.993340 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Mar 17 15:52:56 crc kubenswrapper[4767]: W0317 15:52:56.043191 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2389106_04cc_4871_b544_f8a677a6405a.slice/crio-247a327d3fecd0d0c7f4889439b1e798badca8d5e9c93405f4c5c292b049465a WatchSource:0}: Error finding container 247a327d3fecd0d0c7f4889439b1e798badca8d5e9c93405f4c5c292b049465a: Status 404 returned error can't find the container with id 247a327d3fecd0d0c7f4889439b1e798badca8d5e9c93405f4c5c292b049465a Mar 17 15:52:56 crc kubenswrapper[4767]: I0317 15:52:56.076758 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Mar 17 15:52:56 crc kubenswrapper[4767]: W0317 15:52:56.087050 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27c143b3_0f5b_4265_a2e2_d4302a2f3c70.slice/crio-d556ef39516b269cacf344d95521749c0dbaeaf4dbfa331ecd94c62f5d0f6ef6 WatchSource:0}: Error finding container d556ef39516b269cacf344d95521749c0dbaeaf4dbfa331ecd94c62f5d0f6ef6: Status 404 returned error can't find the container with id d556ef39516b269cacf344d95521749c0dbaeaf4dbfa331ecd94c62f5d0f6ef6 Mar 17 15:52:56 crc kubenswrapper[4767]: I0317 15:52:56.681070 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v75gc" event={"ID":"04d79068-7ef7-42f8-81e5-c5211ec20c12","Type":"ContainerStarted","Data":"8d625c278bf1af3d979eaae8c164c62d63a64c5e76c0336af33ee1abc9db59ad"} Mar 17 15:52:56 crc kubenswrapper[4767]: I0317 15:52:56.683655 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"a2389106-04cc-4871-b544-f8a677a6405a","Type":"ContainerStarted","Data":"247a327d3fecd0d0c7f4889439b1e798badca8d5e9c93405f4c5c292b049465a"} Mar 17 15:52:56 crc kubenswrapper[4767]: I0317 15:52:56.685973 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"27c143b3-0f5b-4265-a2e2-d4302a2f3c70","Type":"ContainerStarted","Data":"d556ef39516b269cacf344d95521749c0dbaeaf4dbfa331ecd94c62f5d0f6ef6"} Mar 17 15:52:56 crc kubenswrapper[4767]: I0317 15:52:56.687909 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" event={"ID":"560edfa4-8a88-4c9b-8b31-e61f93050c15","Type":"ContainerStarted","Data":"6d5a321034bbafa1f08f00aad3a069fa00f4d5c062437199366f23491094c9e1"} Mar 17 15:52:56 crc kubenswrapper[4767]: I0317 15:52:56.690806 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" event={"ID":"9c622cae-f125-4675-9bb5-d2dd999edfa3","Type":"ContainerStarted","Data":"59140645b369bfa9b8e1a3f3dc867272fafcf17466ec5cf7574308f06bc7b26a"} Mar 17 15:52:59 crc kubenswrapper[4767]: I0317 15:52:59.956854 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:52:59 crc kubenswrapper[4767]: I0317 15:52:59.958098 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:53:00 crc kubenswrapper[4767]: I0317 15:53:00.039140 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:53:01 crc kubenswrapper[4767]: I0317 15:53:01.174334 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:53:01 crc kubenswrapper[4767]: I0317 15:53:01.246777 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v75gc"] Mar 17 15:53:01 crc kubenswrapper[4767]: I0317 15:53:01.962795 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"27c143b3-0f5b-4265-a2e2-d4302a2f3c70","Type":"ContainerStarted","Data":"1208416c3424b881ff78fddcc4a361612dc9776e5aeb5af14e619c6424524536"} Mar 17 15:53:01 crc kubenswrapper[4767]: I0317 15:53:01.962911 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:53:01 crc kubenswrapper[4767]: I0317 15:53:01.970385 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" event={"ID":"560edfa4-8a88-4c9b-8b31-e61f93050c15","Type":"ContainerStarted","Data":"d5e376cd8de68a3d0dddf27c7448a2bdf03024562845861b0ff5e7c2a83cfc95"} Mar 17 15:53:01 crc kubenswrapper[4767]: I0317 15:53:01.978519 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" event={"ID":"28ac74bb-3f68-478d-8f50-acf61b2e8223","Type":"ContainerStarted","Data":"c855e14b4c2f3ef66b1ccaede940688b3889efc1a3fea2f3a2a1869f8937960e"} Mar 17 15:53:01 crc kubenswrapper[4767]: I0317 15:53:01.978878 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:53:01 crc kubenswrapper[4767]: I0317 15:53:01.991870 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:53:01 crc kubenswrapper[4767]: I0317 15:53:01.995595 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.673725036 podStartE2EDuration="8.995562027s" podCreationTimestamp="2026-03-17 15:52:53 +0000 UTC" firstStartedPulling="2026-03-17 15:52:56.09651765 +0000 UTC m=+967.509833697" lastFinishedPulling="2026-03-17 15:53:01.418354641 +0000 UTC m=+972.831670688" observedRunningTime="2026-03-17 15:53:01.981145547 +0000 UTC m=+973.394461614" watchObservedRunningTime="2026-03-17 15:53:01.995562027 +0000 UTC m=+973.408878074" Mar 17 15:53:01 crc kubenswrapper[4767]: I0317 15:53:01.996861 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" event={"ID":"66451b46-d0f5-4037-aa4d-abf0143b2ef8","Type":"ContainerStarted","Data":"be8020adc238ffb94f7686545127c1ff4848f129cb7477b9070633c20a4c4eab"} Mar 17 15:53:02 crc kubenswrapper[4767]: I0317 15:53:02.033481 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" podStartSLOduration=3.471792532 podStartE2EDuration="10.03344725s" podCreationTimestamp="2026-03-17 15:52:52 +0000 UTC" firstStartedPulling="2026-03-17 15:52:54.923785046 +0000 UTC m=+966.337101093" lastFinishedPulling="2026-03-17 15:53:01.485439764 +0000 UTC m=+972.898755811" observedRunningTime="2026-03-17 15:53:02.012556472 +0000 UTC m=+973.425872529" watchObservedRunningTime="2026-03-17 15:53:02.03344725 +0000 UTC m=+973.446763297" Mar 17 15:53:02 crc kubenswrapper[4767]: I0317 15:53:02.116610 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" podStartSLOduration=2.494853273 podStartE2EDuration="9.11657075s" podCreationTimestamp="2026-03-17 15:52:53 +0000 UTC" firstStartedPulling="2026-03-17 15:52:54.938449593 +0000 UTC m=+966.351765640" lastFinishedPulling="2026-03-17 15:53:01.56016708 +0000 UTC m=+972.973483117" observedRunningTime="2026-03-17 15:53:02.111851983 +0000 UTC m=+973.525168050" watchObservedRunningTime="2026-03-17 15:53:02.11657075 +0000 UTC m=+973.529886827" Mar 17 15:53:02 crc kubenswrapper[4767]: I0317 15:53:02.143832 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" podStartSLOduration=3.043468037 podStartE2EDuration="9.143795743s" podCreationTimestamp="2026-03-17 15:52:53 +0000 UTC" firstStartedPulling="2026-03-17 15:52:55.4511263 +0000 UTC m=+966.864442347" lastFinishedPulling="2026-03-17 15:53:01.551454006 +0000 UTC m=+972.964770053" observedRunningTime="2026-03-17 15:53:02.135708858 +0000 UTC m=+973.549024905" watchObservedRunningTime="2026-03-17 15:53:02.143795743 +0000 UTC m=+973.557111800" Mar 17 15:53:03 crc kubenswrapper[4767]: I0317 15:53:03.084375 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"a2389106-04cc-4871-b544-f8a677a6405a","Type":"ContainerStarted","Data":"f6e25ef56a9cefef134126498bbd0851ed638c21229507170eb4477bed12f436"} Mar 17 15:53:03 crc kubenswrapper[4767]: I0317 15:53:03.085120 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:53:03 crc kubenswrapper[4767]: I0317 15:53:03.094972 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"89ce8746-1721-4dec-9742-0abc2f498de3","Type":"ContainerStarted","Data":"6f9461e3b5ea3d9251bcb32fa2e70a5d0fc9b9aa2fabc748ae5e22ded9038338"} Mar 17 15:53:03 crc kubenswrapper[4767]: I0317 15:53:03.113030 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" event={"ID":"ce1ebb9b-1558-4190-9e47-4ef6ba3160da","Type":"ContainerStarted","Data":"3bc3aa79c0fbe3a7eb1cd9e57ff876ad92b7a5022493d4507257fc847f3da6c0"} Mar 17 15:53:03 crc kubenswrapper[4767]: I0317 15:53:03.128423 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" event={"ID":"9c622cae-f125-4675-9bb5-d2dd999edfa3","Type":"ContainerStarted","Data":"a874d1d29b42b23e2c6692bd01fa8146578c8cf9a4b7d1cc05839d7a08461c22"} Mar 17 15:53:03 crc kubenswrapper[4767]: I0317 15:53:03.128394 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v75gc" podUID="04d79068-7ef7-42f8-81e5-c5211ec20c12" containerName="registry-server" containerID="cri-o://8d625c278bf1af3d979eaae8c164c62d63a64c5e76c0336af33ee1abc9db59ad" gracePeriod=2 Mar 17 15:53:03 crc kubenswrapper[4767]: I0317 15:53:03.129319 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:53:03 crc kubenswrapper[4767]: I0317 15:53:03.149909 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=4.609371368 podStartE2EDuration="10.149875545s" podCreationTimestamp="2026-03-17 15:52:53 +0000 UTC" firstStartedPulling="2026-03-17 15:52:56.046635058 +0000 UTC m=+967.459951105" lastFinishedPulling="2026-03-17 15:53:01.587139235 +0000 UTC m=+973.000455282" observedRunningTime="2026-03-17 15:53:03.107557573 +0000 UTC m=+974.520873640" watchObservedRunningTime="2026-03-17 15:53:03.149875545 +0000 UTC m=+974.563191592" Mar 17 15:53:03 crc kubenswrapper[4767]: I0317 15:53:03.166468 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=4.15543223 podStartE2EDuration="10.166437527s" podCreationTimestamp="2026-03-17 15:52:53 +0000 UTC" firstStartedPulling="2026-03-17 15:52:55.549090031 +0000 UTC m=+966.962406088" lastFinishedPulling="2026-03-17 15:53:01.560095348 +0000 UTC m=+972.973411385" observedRunningTime="2026-03-17 15:53:03.131989874 +0000 UTC m=+974.545305941" watchObservedRunningTime="2026-03-17 15:53:03.166437527 +0000 UTC m=+974.579753574" Mar 17 15:53:04 crc kubenswrapper[4767]: I0317 15:53:04.163453 4767 generic.go:334] "Generic (PLEG): container finished" podID="04d79068-7ef7-42f8-81e5-c5211ec20c12" containerID="8d625c278bf1af3d979eaae8c164c62d63a64c5e76c0336af33ee1abc9db59ad" exitCode=0 Mar 17 15:53:04 crc kubenswrapper[4767]: I0317 15:53:04.163507 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v75gc" event={"ID":"04d79068-7ef7-42f8-81e5-c5211ec20c12","Type":"ContainerDied","Data":"8d625c278bf1af3d979eaae8c164c62d63a64c5e76c0336af33ee1abc9db59ad"} Mar 17 15:53:04 crc kubenswrapper[4767]: I0317 15:53:04.291847 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:53:04 crc kubenswrapper[4767]: I0317 15:53:04.895897 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.022633 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-utilities\") pod \"04d79068-7ef7-42f8-81e5-c5211ec20c12\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.022714 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd4c2\" (UniqueName: \"kubernetes.io/projected/04d79068-7ef7-42f8-81e5-c5211ec20c12-kube-api-access-gd4c2\") pod \"04d79068-7ef7-42f8-81e5-c5211ec20c12\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.022834 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-catalog-content\") pod \"04d79068-7ef7-42f8-81e5-c5211ec20c12\" (UID: \"04d79068-7ef7-42f8-81e5-c5211ec20c12\") " Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.024218 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-utilities" (OuterVolumeSpecName: "utilities") pod "04d79068-7ef7-42f8-81e5-c5211ec20c12" (UID: "04d79068-7ef7-42f8-81e5-c5211ec20c12"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.038746 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04d79068-7ef7-42f8-81e5-c5211ec20c12-kube-api-access-gd4c2" (OuterVolumeSpecName: "kube-api-access-gd4c2") pod "04d79068-7ef7-42f8-81e5-c5211ec20c12" (UID: "04d79068-7ef7-42f8-81e5-c5211ec20c12"). InnerVolumeSpecName "kube-api-access-gd4c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.323621 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.323695 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd4c2\" (UniqueName: \"kubernetes.io/projected/04d79068-7ef7-42f8-81e5-c5211ec20c12-kube-api-access-gd4c2\") on node \"crc\" DevicePath \"\"" Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.366504 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v75gc" Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.371022 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04d79068-7ef7-42f8-81e5-c5211ec20c12" (UID: "04d79068-7ef7-42f8-81e5-c5211ec20c12"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.374222 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v75gc" event={"ID":"04d79068-7ef7-42f8-81e5-c5211ec20c12","Type":"ContainerDied","Data":"2e9cf3eb3d4a8ca2977bfdf7602b67857ddcad2fad149b2567630442a2559864"} Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.374284 4767 scope.go:117] "RemoveContainer" containerID="8d625c278bf1af3d979eaae8c164c62d63a64c5e76c0336af33ee1abc9db59ad" Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.399148 4767 scope.go:117] "RemoveContainer" containerID="4c91a5e32f216332ae10702575de4cfcfe0ed53ce37a9c1c52193afe8298dd32" Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.433791 4767 scope.go:117] "RemoveContainer" containerID="84046a4dbc6d48fb9bbf0fc1dd851ebd8844dbef28561f13c67cce6473db8870" Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.437065 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04d79068-7ef7-42f8-81e5-c5211ec20c12-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.694069 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v75gc"] Mar 17 15:53:05 crc kubenswrapper[4767]: I0317 15:53:05.700614 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v75gc"] Mar 17 15:53:06 crc kubenswrapper[4767]: I0317 15:53:06.376674 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" event={"ID":"560edfa4-8a88-4c9b-8b31-e61f93050c15","Type":"ContainerStarted","Data":"a1bb42f9568f0016b3faf38f6522782f5838daac9d1d8a69dcada82fa7207b97"} Mar 17 15:53:06 crc kubenswrapper[4767]: I0317 15:53:06.377343 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:53:06 crc kubenswrapper[4767]: I0317 15:53:06.382030 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" event={"ID":"9c622cae-f125-4675-9bb5-d2dd999edfa3","Type":"ContainerStarted","Data":"17a6f3d39868811621ffcfbf05f957f7ff878993657e976173932bb7bd919e9c"} Mar 17 15:53:06 crc kubenswrapper[4767]: I0317 15:53:06.382307 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:53:06 crc kubenswrapper[4767]: I0317 15:53:06.382358 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:53:06 crc kubenswrapper[4767]: I0317 15:53:06.390720 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:53:06 crc kubenswrapper[4767]: I0317 15:53:06.391538 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:53:06 crc kubenswrapper[4767]: I0317 15:53:06.394097 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" Mar 17 15:53:06 crc kubenswrapper[4767]: I0317 15:53:06.406263 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podStartSLOduration=2.250554035 podStartE2EDuration="12.406234817s" podCreationTimestamp="2026-03-17 15:52:54 +0000 UTC" firstStartedPulling="2026-03-17 15:52:55.923434931 +0000 UTC m=+967.336750978" lastFinishedPulling="2026-03-17 15:53:06.079115713 +0000 UTC m=+977.492431760" observedRunningTime="2026-03-17 15:53:06.404257399 +0000 UTC m=+977.817573456" watchObservedRunningTime="2026-03-17 15:53:06.406234817 +0000 UTC m=+977.819550854" Mar 17 15:53:06 crc kubenswrapper[4767]: I0317 15:53:06.431129 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podStartSLOduration=2.152465159 podStartE2EDuration="12.431090661s" podCreationTimestamp="2026-03-17 15:52:54 +0000 UTC" firstStartedPulling="2026-03-17 15:52:55.810960536 +0000 UTC m=+967.224276583" lastFinishedPulling="2026-03-17 15:53:06.089586038 +0000 UTC m=+977.502902085" observedRunningTime="2026-03-17 15:53:06.424940671 +0000 UTC m=+977.838256728" watchObservedRunningTime="2026-03-17 15:53:06.431090661 +0000 UTC m=+977.844406718" Mar 17 15:53:07 crc kubenswrapper[4767]: I0317 15:53:07.364757 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04d79068-7ef7-42f8-81e5-c5211ec20c12" path="/var/lib/kubelet/pods/04d79068-7ef7-42f8-81e5-c5211ec20c12/volumes" Mar 17 15:53:07 crc kubenswrapper[4767]: I0317 15:53:07.393454 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:53:07 crc kubenswrapper[4767]: I0317 15:53:07.402709 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" Mar 17 15:53:23 crc kubenswrapper[4767]: I0317 15:53:23.495852 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 15:53:23 crc kubenswrapper[4767]: I0317 15:53:23.664744 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 15:53:24 crc kubenswrapper[4767]: I0317 15:53:24.393547 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" Mar 17 15:53:25 crc kubenswrapper[4767]: I0317 15:53:25.088396 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Mar 17 15:53:25 crc kubenswrapper[4767]: I0317 15:53:25.169068 4767 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Mar 17 15:53:25 crc kubenswrapper[4767]: I0317 15:53:25.169151 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a2389106-04cc-4871-b544-f8a677a6405a" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 15:53:25 crc kubenswrapper[4767]: I0317 15:53:25.186506 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Mar 17 15:53:34 crc kubenswrapper[4767]: I0317 15:53:34.166869 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:53:34 crc kubenswrapper[4767]: I0317 15:53:34.167820 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:53:35 crc kubenswrapper[4767]: I0317 15:53:35.158319 4767 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Mar 17 15:53:35 crc kubenswrapper[4767]: I0317 15:53:35.158947 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a2389106-04cc-4871-b544-f8a677a6405a" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.568549 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-grt4r"] Mar 17 15:53:42 crc kubenswrapper[4767]: E0317 15:53:42.569893 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d79068-7ef7-42f8-81e5-c5211ec20c12" containerName="registry-server" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.569915 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d79068-7ef7-42f8-81e5-c5211ec20c12" containerName="registry-server" Mar 17 15:53:42 crc kubenswrapper[4767]: E0317 15:53:42.569933 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d79068-7ef7-42f8-81e5-c5211ec20c12" containerName="extract-content" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.569942 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d79068-7ef7-42f8-81e5-c5211ec20c12" containerName="extract-content" Mar 17 15:53:42 crc kubenswrapper[4767]: E0317 15:53:42.569959 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d79068-7ef7-42f8-81e5-c5211ec20c12" containerName="extract-utilities" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.569965 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d79068-7ef7-42f8-81e5-c5211ec20c12" containerName="extract-utilities" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.570193 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="04d79068-7ef7-42f8-81e5-c5211ec20c12" containerName="registry-server" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.571581 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.598640 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-grt4r"] Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.688383 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-utilities\") pod \"redhat-marketplace-grt4r\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.688571 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rhdh\" (UniqueName: \"kubernetes.io/projected/e5bbc752-2331-47f5-85ee-46c8acc27793-kube-api-access-2rhdh\") pod \"redhat-marketplace-grt4r\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.688649 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-catalog-content\") pod \"redhat-marketplace-grt4r\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.790419 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rhdh\" (UniqueName: \"kubernetes.io/projected/e5bbc752-2331-47f5-85ee-46c8acc27793-kube-api-access-2rhdh\") pod \"redhat-marketplace-grt4r\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.790858 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-catalog-content\") pod \"redhat-marketplace-grt4r\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.790975 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-utilities\") pod \"redhat-marketplace-grt4r\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.791586 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-utilities\") pod \"redhat-marketplace-grt4r\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.791915 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-catalog-content\") pod \"redhat-marketplace-grt4r\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.818801 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rhdh\" (UniqueName: \"kubernetes.io/projected/e5bbc752-2331-47f5-85ee-46c8acc27793-kube-api-access-2rhdh\") pod \"redhat-marketplace-grt4r\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:42 crc kubenswrapper[4767]: I0317 15:53:42.894639 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:43 crc kubenswrapper[4767]: I0317 15:53:43.409571 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-grt4r"] Mar 17 15:53:43 crc kubenswrapper[4767]: I0317 15:53:43.718957 4767 generic.go:334] "Generic (PLEG): container finished" podID="e5bbc752-2331-47f5-85ee-46c8acc27793" containerID="04078f0953da48fed74ae73188214b60b64330da31dae82671a480f1450cfce3" exitCode=0 Mar 17 15:53:43 crc kubenswrapper[4767]: I0317 15:53:43.719141 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grt4r" event={"ID":"e5bbc752-2331-47f5-85ee-46c8acc27793","Type":"ContainerDied","Data":"04078f0953da48fed74ae73188214b60b64330da31dae82671a480f1450cfce3"} Mar 17 15:53:43 crc kubenswrapper[4767]: I0317 15:53:43.719507 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grt4r" event={"ID":"e5bbc752-2331-47f5-85ee-46c8acc27793","Type":"ContainerStarted","Data":"79ea1aba63d5981cd6625755a3963686722c5763dca357a38b87a8f59f0fa027"} Mar 17 15:53:44 crc kubenswrapper[4767]: I0317 15:53:44.733368 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grt4r" event={"ID":"e5bbc752-2331-47f5-85ee-46c8acc27793","Type":"ContainerStarted","Data":"8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0"} Mar 17 15:53:45 crc kubenswrapper[4767]: I0317 15:53:45.250823 4767 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Mar 17 15:53:45 crc kubenswrapper[4767]: I0317 15:53:45.250949 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a2389106-04cc-4871-b544-f8a677a6405a" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 15:53:45 crc kubenswrapper[4767]: I0317 15:53:45.746210 4767 generic.go:334] "Generic (PLEG): container finished" podID="e5bbc752-2331-47f5-85ee-46c8acc27793" containerID="8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0" exitCode=0 Mar 17 15:53:45 crc kubenswrapper[4767]: I0317 15:53:45.746311 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grt4r" event={"ID":"e5bbc752-2331-47f5-85ee-46c8acc27793","Type":"ContainerDied","Data":"8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0"} Mar 17 15:53:46 crc kubenswrapper[4767]: I0317 15:53:46.760338 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grt4r" event={"ID":"e5bbc752-2331-47f5-85ee-46c8acc27793","Type":"ContainerStarted","Data":"767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e"} Mar 17 15:53:46 crc kubenswrapper[4767]: I0317 15:53:46.785204 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-grt4r" podStartSLOduration=2.1944356799999998 podStartE2EDuration="4.785149511s" podCreationTimestamp="2026-03-17 15:53:42 +0000 UTC" firstStartedPulling="2026-03-17 15:53:43.723403855 +0000 UTC m=+1015.136719902" lastFinishedPulling="2026-03-17 15:53:46.314117686 +0000 UTC m=+1017.727433733" observedRunningTime="2026-03-17 15:53:46.781739961 +0000 UTC m=+1018.195056008" watchObservedRunningTime="2026-03-17 15:53:46.785149511 +0000 UTC m=+1018.198465548" Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.254784 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lzmr8"] Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.258109 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.294299 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lzmr8"] Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.373109 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-utilities\") pod \"certified-operators-lzmr8\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.373246 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-catalog-content\") pod \"certified-operators-lzmr8\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.373400 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jt92\" (UniqueName: \"kubernetes.io/projected/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-kube-api-access-9jt92\") pod \"certified-operators-lzmr8\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.474901 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jt92\" (UniqueName: \"kubernetes.io/projected/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-kube-api-access-9jt92\") pod \"certified-operators-lzmr8\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.475078 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-utilities\") pod \"certified-operators-lzmr8\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.475104 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-catalog-content\") pod \"certified-operators-lzmr8\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.475663 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-catalog-content\") pod \"certified-operators-lzmr8\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.475856 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-utilities\") pod \"certified-operators-lzmr8\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.505367 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jt92\" (UniqueName: \"kubernetes.io/projected/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-kube-api-access-9jt92\") pod \"certified-operators-lzmr8\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:53:50 crc kubenswrapper[4767]: I0317 15:53:50.586669 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:53:51 crc kubenswrapper[4767]: I0317 15:53:51.299408 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lzmr8"] Mar 17 15:53:51 crc kubenswrapper[4767]: W0317 15:53:51.302881 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeac2ecc1_ad2c_4d3f_98b7_a8b235d8c78d.slice/crio-3e945350f0be4f51e1dd0fb232f2716c95c08830eb8dfd81505267ebe8458624 WatchSource:0}: Error finding container 3e945350f0be4f51e1dd0fb232f2716c95c08830eb8dfd81505267ebe8458624: Status 404 returned error can't find the container with id 3e945350f0be4f51e1dd0fb232f2716c95c08830eb8dfd81505267ebe8458624 Mar 17 15:53:52 crc kubenswrapper[4767]: I0317 15:53:52.079988 4767 generic.go:334] "Generic (PLEG): container finished" podID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" containerID="8c0df96de2952a523d6fa3b4bf1681e6e0b5068ecee9e203390fffb43ffe10cc" exitCode=0 Mar 17 15:53:52 crc kubenswrapper[4767]: I0317 15:53:52.080389 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzmr8" event={"ID":"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d","Type":"ContainerDied","Data":"8c0df96de2952a523d6fa3b4bf1681e6e0b5068ecee9e203390fffb43ffe10cc"} Mar 17 15:53:52 crc kubenswrapper[4767]: I0317 15:53:52.080479 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzmr8" event={"ID":"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d","Type":"ContainerStarted","Data":"3e945350f0be4f51e1dd0fb232f2716c95c08830eb8dfd81505267ebe8458624"} Mar 17 15:53:52 crc kubenswrapper[4767]: I0317 15:53:52.895586 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:52 crc kubenswrapper[4767]: I0317 15:53:52.896199 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:52 crc kubenswrapper[4767]: I0317 15:53:52.963881 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:53 crc kubenswrapper[4767]: I0317 15:53:53.092666 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzmr8" event={"ID":"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d","Type":"ContainerStarted","Data":"9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e"} Mar 17 15:53:53 crc kubenswrapper[4767]: I0317 15:53:53.154798 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:54 crc kubenswrapper[4767]: I0317 15:53:54.104591 4767 generic.go:334] "Generic (PLEG): container finished" podID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" containerID="9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e" exitCode=0 Mar 17 15:53:54 crc kubenswrapper[4767]: I0317 15:53:54.104678 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzmr8" event={"ID":"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d","Type":"ContainerDied","Data":"9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e"} Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.116677 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzmr8" event={"ID":"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d","Type":"ContainerStarted","Data":"4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac"} Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.148346 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lzmr8" podStartSLOduration=2.726077778 podStartE2EDuration="5.148322664s" podCreationTimestamp="2026-03-17 15:53:50 +0000 UTC" firstStartedPulling="2026-03-17 15:53:52.082127648 +0000 UTC m=+1023.495443695" lastFinishedPulling="2026-03-17 15:53:54.504372494 +0000 UTC m=+1025.917688581" observedRunningTime="2026-03-17 15:53:55.146209482 +0000 UTC m=+1026.559525539" watchObservedRunningTime="2026-03-17 15:53:55.148322664 +0000 UTC m=+1026.561638711" Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.163568 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.331414 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-grt4r"] Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.332316 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-grt4r" podUID="e5bbc752-2331-47f5-85ee-46c8acc27793" containerName="registry-server" containerID="cri-o://767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e" gracePeriod=2 Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.738314 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.890628 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-catalog-content\") pod \"e5bbc752-2331-47f5-85ee-46c8acc27793\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.890716 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rhdh\" (UniqueName: \"kubernetes.io/projected/e5bbc752-2331-47f5-85ee-46c8acc27793-kube-api-access-2rhdh\") pod \"e5bbc752-2331-47f5-85ee-46c8acc27793\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.890879 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-utilities\") pod \"e5bbc752-2331-47f5-85ee-46c8acc27793\" (UID: \"e5bbc752-2331-47f5-85ee-46c8acc27793\") " Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.891968 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-utilities" (OuterVolumeSpecName: "utilities") pod "e5bbc752-2331-47f5-85ee-46c8acc27793" (UID: "e5bbc752-2331-47f5-85ee-46c8acc27793"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.901298 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5bbc752-2331-47f5-85ee-46c8acc27793-kube-api-access-2rhdh" (OuterVolumeSpecName: "kube-api-access-2rhdh") pod "e5bbc752-2331-47f5-85ee-46c8acc27793" (UID: "e5bbc752-2331-47f5-85ee-46c8acc27793"). InnerVolumeSpecName "kube-api-access-2rhdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.920008 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5bbc752-2331-47f5-85ee-46c8acc27793" (UID: "e5bbc752-2331-47f5-85ee-46c8acc27793"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.993280 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.993333 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5bbc752-2331-47f5-85ee-46c8acc27793-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:53:55 crc kubenswrapper[4767]: I0317 15:53:55.993353 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rhdh\" (UniqueName: \"kubernetes.io/projected/e5bbc752-2331-47f5-85ee-46c8acc27793-kube-api-access-2rhdh\") on node \"crc\" DevicePath \"\"" Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.127582 4767 generic.go:334] "Generic (PLEG): container finished" podID="e5bbc752-2331-47f5-85ee-46c8acc27793" containerID="767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e" exitCode=0 Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.127668 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-grt4r" Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.127702 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grt4r" event={"ID":"e5bbc752-2331-47f5-85ee-46c8acc27793","Type":"ContainerDied","Data":"767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e"} Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.127774 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-grt4r" event={"ID":"e5bbc752-2331-47f5-85ee-46c8acc27793","Type":"ContainerDied","Data":"79ea1aba63d5981cd6625755a3963686722c5763dca357a38b87a8f59f0fa027"} Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.127808 4767 scope.go:117] "RemoveContainer" containerID="767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e" Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.151382 4767 scope.go:117] "RemoveContainer" containerID="8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0" Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.181291 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-grt4r"] Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.187316 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-grt4r"] Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.204692 4767 scope.go:117] "RemoveContainer" containerID="04078f0953da48fed74ae73188214b60b64330da31dae82671a480f1450cfce3" Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.236470 4767 scope.go:117] "RemoveContainer" containerID="767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e" Mar 17 15:53:56 crc kubenswrapper[4767]: E0317 15:53:56.238585 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e\": container with ID starting with 767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e not found: ID does not exist" containerID="767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e" Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.238648 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e"} err="failed to get container status \"767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e\": rpc error: code = NotFound desc = could not find container \"767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e\": container with ID starting with 767be9fb8b0b5bc07c73dace619633de9621f294d60c041ea97b543e12c9185e not found: ID does not exist" Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.238688 4767 scope.go:117] "RemoveContainer" containerID="8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0" Mar 17 15:53:56 crc kubenswrapper[4767]: E0317 15:53:56.239498 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0\": container with ID starting with 8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0 not found: ID does not exist" containerID="8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0" Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.239560 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0"} err="failed to get container status \"8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0\": rpc error: code = NotFound desc = could not find container \"8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0\": container with ID starting with 8154769b58f3c451277f096be94b05926f66312b0e1cd7d260070997c48d3ff0 not found: ID does not exist" Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.239602 4767 scope.go:117] "RemoveContainer" containerID="04078f0953da48fed74ae73188214b60b64330da31dae82671a480f1450cfce3" Mar 17 15:53:56 crc kubenswrapper[4767]: E0317 15:53:56.240251 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04078f0953da48fed74ae73188214b60b64330da31dae82671a480f1450cfce3\": container with ID starting with 04078f0953da48fed74ae73188214b60b64330da31dae82671a480f1450cfce3 not found: ID does not exist" containerID="04078f0953da48fed74ae73188214b60b64330da31dae82671a480f1450cfce3" Mar 17 15:53:56 crc kubenswrapper[4767]: I0317 15:53:56.240317 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04078f0953da48fed74ae73188214b60b64330da31dae82671a480f1450cfce3"} err="failed to get container status \"04078f0953da48fed74ae73188214b60b64330da31dae82671a480f1450cfce3\": rpc error: code = NotFound desc = could not find container \"04078f0953da48fed74ae73188214b60b64330da31dae82671a480f1450cfce3\": container with ID starting with 04078f0953da48fed74ae73188214b60b64330da31dae82671a480f1450cfce3 not found: ID does not exist" Mar 17 15:53:57 crc kubenswrapper[4767]: I0317 15:53:57.451321 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5bbc752-2331-47f5-85ee-46c8acc27793" path="/var/lib/kubelet/pods/e5bbc752-2331-47f5-85ee-46c8acc27793/volumes" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.147930 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562714-xlm92"] Mar 17 15:54:00 crc kubenswrapper[4767]: E0317 15:54:00.149050 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5bbc752-2331-47f5-85ee-46c8acc27793" containerName="registry-server" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.149072 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5bbc752-2331-47f5-85ee-46c8acc27793" containerName="registry-server" Mar 17 15:54:00 crc kubenswrapper[4767]: E0317 15:54:00.149087 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5bbc752-2331-47f5-85ee-46c8acc27793" containerName="extract-utilities" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.149095 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5bbc752-2331-47f5-85ee-46c8acc27793" containerName="extract-utilities" Mar 17 15:54:00 crc kubenswrapper[4767]: E0317 15:54:00.149122 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5bbc752-2331-47f5-85ee-46c8acc27793" containerName="extract-content" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.149130 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5bbc752-2331-47f5-85ee-46c8acc27793" containerName="extract-content" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.149324 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5bbc752-2331-47f5-85ee-46c8acc27793" containerName="registry-server" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.150157 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562714-xlm92" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.153505 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.153551 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.153956 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.157445 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562714-xlm92"] Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.289822 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfw56\" (UniqueName: \"kubernetes.io/projected/ae556c0b-e4f1-42ce-98de-97e6241f8b5d-kube-api-access-jfw56\") pod \"auto-csr-approver-29562714-xlm92\" (UID: \"ae556c0b-e4f1-42ce-98de-97e6241f8b5d\") " pod="openshift-infra/auto-csr-approver-29562714-xlm92" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.391883 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfw56\" (UniqueName: \"kubernetes.io/projected/ae556c0b-e4f1-42ce-98de-97e6241f8b5d-kube-api-access-jfw56\") pod \"auto-csr-approver-29562714-xlm92\" (UID: \"ae556c0b-e4f1-42ce-98de-97e6241f8b5d\") " pod="openshift-infra/auto-csr-approver-29562714-xlm92" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.413892 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfw56\" (UniqueName: \"kubernetes.io/projected/ae556c0b-e4f1-42ce-98de-97e6241f8b5d-kube-api-access-jfw56\") pod \"auto-csr-approver-29562714-xlm92\" (UID: \"ae556c0b-e4f1-42ce-98de-97e6241f8b5d\") " pod="openshift-infra/auto-csr-approver-29562714-xlm92" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.483850 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562714-xlm92" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.589664 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.589758 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.713807 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:54:00 crc kubenswrapper[4767]: I0317 15:54:00.965684 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562714-xlm92"] Mar 17 15:54:01 crc kubenswrapper[4767]: I0317 15:54:01.483436 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562714-xlm92" event={"ID":"ae556c0b-e4f1-42ce-98de-97e6241f8b5d","Type":"ContainerStarted","Data":"bb1259a0a7cc45c21b84cdc41f15caf22bba16dee2269bead77cfb565a96786e"} Mar 17 15:54:01 crc kubenswrapper[4767]: I0317 15:54:01.530371 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:54:01 crc kubenswrapper[4767]: I0317 15:54:01.592487 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lzmr8"] Mar 17 15:54:02 crc kubenswrapper[4767]: I0317 15:54:02.493933 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562714-xlm92" event={"ID":"ae556c0b-e4f1-42ce-98de-97e6241f8b5d","Type":"ContainerStarted","Data":"7c3acd80733a58dbc10ddc2eea984975cf074e58e075379c6fb58bf80bc50395"} Mar 17 15:54:02 crc kubenswrapper[4767]: I0317 15:54:02.517202 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562714-xlm92" podStartSLOduration=1.35273896 podStartE2EDuration="2.517153664s" podCreationTimestamp="2026-03-17 15:54:00 +0000 UTC" firstStartedPulling="2026-03-17 15:54:00.979679988 +0000 UTC m=+1032.392996035" lastFinishedPulling="2026-03-17 15:54:02.144094692 +0000 UTC m=+1033.557410739" observedRunningTime="2026-03-17 15:54:02.515064323 +0000 UTC m=+1033.928380380" watchObservedRunningTime="2026-03-17 15:54:02.517153664 +0000 UTC m=+1033.930469711" Mar 17 15:54:03 crc kubenswrapper[4767]: I0317 15:54:03.505834 4767 generic.go:334] "Generic (PLEG): container finished" podID="ae556c0b-e4f1-42ce-98de-97e6241f8b5d" containerID="7c3acd80733a58dbc10ddc2eea984975cf074e58e075379c6fb58bf80bc50395" exitCode=0 Mar 17 15:54:03 crc kubenswrapper[4767]: I0317 15:54:03.505934 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562714-xlm92" event={"ID":"ae556c0b-e4f1-42ce-98de-97e6241f8b5d","Type":"ContainerDied","Data":"7c3acd80733a58dbc10ddc2eea984975cf074e58e075379c6fb58bf80bc50395"} Mar 17 15:54:03 crc kubenswrapper[4767]: I0317 15:54:03.506677 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lzmr8" podUID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" containerName="registry-server" containerID="cri-o://4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac" gracePeriod=2 Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.166927 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.167609 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.428828 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.519021 4767 generic.go:334] "Generic (PLEG): container finished" podID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" containerID="4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac" exitCode=0 Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.519292 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzmr8" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.520111 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzmr8" event={"ID":"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d","Type":"ContainerDied","Data":"4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac"} Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.520149 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzmr8" event={"ID":"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d","Type":"ContainerDied","Data":"3e945350f0be4f51e1dd0fb232f2716c95c08830eb8dfd81505267ebe8458624"} Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.520183 4767 scope.go:117] "RemoveContainer" containerID="4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.544961 4767 scope.go:117] "RemoveContainer" containerID="9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.573569 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-catalog-content\") pod \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.573689 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jt92\" (UniqueName: \"kubernetes.io/projected/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-kube-api-access-9jt92\") pod \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.573739 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-utilities\") pod \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\" (UID: \"eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d\") " Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.576533 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-utilities" (OuterVolumeSpecName: "utilities") pod "eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" (UID: "eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.583049 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-kube-api-access-9jt92" (OuterVolumeSpecName: "kube-api-access-9jt92") pod "eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" (UID: "eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d"). InnerVolumeSpecName "kube-api-access-9jt92". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.598752 4767 scope.go:117] "RemoveContainer" containerID="8c0df96de2952a523d6fa3b4bf1681e6e0b5068ecee9e203390fffb43ffe10cc" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.636476 4767 scope.go:117] "RemoveContainer" containerID="4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.637700 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" (UID: "eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:54:04 crc kubenswrapper[4767]: E0317 15:54:04.637945 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac\": container with ID starting with 4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac not found: ID does not exist" containerID="4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.638004 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac"} err="failed to get container status \"4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac\": rpc error: code = NotFound desc = could not find container \"4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac\": container with ID starting with 4147a24a360b40f04ec84562daa5ddb62e2dc490924be291a616f6ac687887ac not found: ID does not exist" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.638051 4767 scope.go:117] "RemoveContainer" containerID="9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e" Mar 17 15:54:04 crc kubenswrapper[4767]: E0317 15:54:04.638419 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e\": container with ID starting with 9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e not found: ID does not exist" containerID="9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.638442 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e"} err="failed to get container status \"9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e\": rpc error: code = NotFound desc = could not find container \"9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e\": container with ID starting with 9ac43634af1276cd01242da86efe7cb4940629ff3857b76a4a6d16ba1fdf293e not found: ID does not exist" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.638464 4767 scope.go:117] "RemoveContainer" containerID="8c0df96de2952a523d6fa3b4bf1681e6e0b5068ecee9e203390fffb43ffe10cc" Mar 17 15:54:04 crc kubenswrapper[4767]: E0317 15:54:04.638687 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c0df96de2952a523d6fa3b4bf1681e6e0b5068ecee9e203390fffb43ffe10cc\": container with ID starting with 8c0df96de2952a523d6fa3b4bf1681e6e0b5068ecee9e203390fffb43ffe10cc not found: ID does not exist" containerID="8c0df96de2952a523d6fa3b4bf1681e6e0b5068ecee9e203390fffb43ffe10cc" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.638720 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c0df96de2952a523d6fa3b4bf1681e6e0b5068ecee9e203390fffb43ffe10cc"} err="failed to get container status \"8c0df96de2952a523d6fa3b4bf1681e6e0b5068ecee9e203390fffb43ffe10cc\": rpc error: code = NotFound desc = could not find container \"8c0df96de2952a523d6fa3b4bf1681e6e0b5068ecee9e203390fffb43ffe10cc\": container with ID starting with 8c0df96de2952a523d6fa3b4bf1681e6e0b5068ecee9e203390fffb43ffe10cc not found: ID does not exist" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.676049 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jt92\" (UniqueName: \"kubernetes.io/projected/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-kube-api-access-9jt92\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.676109 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.676121 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.844832 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562714-xlm92" Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.860591 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lzmr8"] Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.867594 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lzmr8"] Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.982551 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfw56\" (UniqueName: \"kubernetes.io/projected/ae556c0b-e4f1-42ce-98de-97e6241f8b5d-kube-api-access-jfw56\") pod \"ae556c0b-e4f1-42ce-98de-97e6241f8b5d\" (UID: \"ae556c0b-e4f1-42ce-98de-97e6241f8b5d\") " Mar 17 15:54:04 crc kubenswrapper[4767]: I0317 15:54:04.989526 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae556c0b-e4f1-42ce-98de-97e6241f8b5d-kube-api-access-jfw56" (OuterVolumeSpecName: "kube-api-access-jfw56") pod "ae556c0b-e4f1-42ce-98de-97e6241f8b5d" (UID: "ae556c0b-e4f1-42ce-98de-97e6241f8b5d"). InnerVolumeSpecName "kube-api-access-jfw56". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:54:05 crc kubenswrapper[4767]: I0317 15:54:05.085801 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfw56\" (UniqueName: \"kubernetes.io/projected/ae556c0b-e4f1-42ce-98de-97e6241f8b5d-kube-api-access-jfw56\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:05 crc kubenswrapper[4767]: I0317 15:54:05.366631 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" path="/var/lib/kubelet/pods/eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d/volumes" Mar 17 15:54:05 crc kubenswrapper[4767]: I0317 15:54:05.534146 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562714-xlm92" event={"ID":"ae556c0b-e4f1-42ce-98de-97e6241f8b5d","Type":"ContainerDied","Data":"bb1259a0a7cc45c21b84cdc41f15caf22bba16dee2269bead77cfb565a96786e"} Mar 17 15:54:05 crc kubenswrapper[4767]: I0317 15:54:05.535184 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb1259a0a7cc45c21b84cdc41f15caf22bba16dee2269bead77cfb565a96786e" Mar 17 15:54:05 crc kubenswrapper[4767]: I0317 15:54:05.534474 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562714-xlm92" Mar 17 15:54:05 crc kubenswrapper[4767]: I0317 15:54:05.660335 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562708-ghrs9"] Mar 17 15:54:05 crc kubenswrapper[4767]: I0317 15:54:05.665883 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562708-ghrs9"] Mar 17 15:54:07 crc kubenswrapper[4767]: I0317 15:54:07.365194 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a97b7496-d795-4914-b980-857e83f3f7da" path="/var/lib/kubelet/pods/a97b7496-d795-4914-b980-857e83f3f7da/volumes" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.233207 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-gdslt"] Mar 17 15:54:11 crc kubenswrapper[4767]: E0317 15:54:11.234467 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" containerName="registry-server" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.234512 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" containerName="registry-server" Mar 17 15:54:11 crc kubenswrapper[4767]: E0317 15:54:11.234539 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae556c0b-e4f1-42ce-98de-97e6241f8b5d" containerName="oc" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.234562 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae556c0b-e4f1-42ce-98de-97e6241f8b5d" containerName="oc" Mar 17 15:54:11 crc kubenswrapper[4767]: E0317 15:54:11.234575 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" containerName="extract-content" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.234582 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" containerName="extract-content" Mar 17 15:54:11 crc kubenswrapper[4767]: E0317 15:54:11.234600 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" containerName="extract-utilities" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.234606 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" containerName="extract-utilities" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.234785 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae556c0b-e4f1-42ce-98de-97e6241f8b5d" containerName="oc" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.234803 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="eac2ecc1-ad2c-4d3f-98b7-a8b235d8c78d" containerName="registry-server" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.235537 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.237930 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.238007 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.238581 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.239984 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-n2c2q" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.240286 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.246105 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.323732 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-gdslt"] Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.407060 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-gdslt"] Mar 17 15:54:11 crc kubenswrapper[4767]: E0317 15:54:11.409252 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-h8cdj metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-gdslt" podUID="101dce53-f7ef-4b92-9c0b-c587edb60b9a" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.418426 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-entrypoint\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.418934 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-token\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.419063 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/101dce53-f7ef-4b92-9c0b-c587edb60b9a-datadir\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.419192 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-trusted-ca\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.419308 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/101dce53-f7ef-4b92-9c0b-c587edb60b9a-tmp\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.419442 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.419573 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-sa-token\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.421206 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-metrics\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.421404 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-syslog-receiver\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.421529 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config-openshift-service-cacrt\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.421682 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8cdj\" (UniqueName: \"kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-kube-api-access-h8cdj\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.524069 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-metrics\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.524156 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-syslog-receiver\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.524273 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config-openshift-service-cacrt\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.524313 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8cdj\" (UniqueName: \"kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-kube-api-access-h8cdj\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.524348 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-entrypoint\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.524371 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-token\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.524395 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/101dce53-f7ef-4b92-9c0b-c587edb60b9a-datadir\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.524442 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-trusted-ca\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.524471 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/101dce53-f7ef-4b92-9c0b-c587edb60b9a-tmp\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.524500 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.524530 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-sa-token\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.525025 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/101dce53-f7ef-4b92-9c0b-c587edb60b9a-datadir\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.526465 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.526676 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-trusted-ca\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.527682 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config-openshift-service-cacrt\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.528337 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-entrypoint\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.531626 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/101dce53-f7ef-4b92-9c0b-c587edb60b9a-tmp\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.532098 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-metrics\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.534981 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-syslog-receiver\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.539304 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-token\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.545776 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8cdj\" (UniqueName: \"kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-kube-api-access-h8cdj\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.548996 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-sa-token\") pod \"collector-gdslt\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.584704 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.604302 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-gdslt" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.625778 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-token\") pod \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.625863 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-syslog-receiver\") pod \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.625916 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config\") pod \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.626001 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8cdj\" (UniqueName: \"kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-kube-api-access-h8cdj\") pod \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.626122 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-metrics\") pod \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.626157 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-entrypoint\") pod \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.626217 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-sa-token\") pod \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.626296 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config-openshift-service-cacrt\") pod \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.626341 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/101dce53-f7ef-4b92-9c0b-c587edb60b9a-datadir\") pod \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.626391 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/101dce53-f7ef-4b92-9c0b-c587edb60b9a-tmp\") pod \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.626427 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-trusted-ca\") pod \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\" (UID: \"101dce53-f7ef-4b92-9c0b-c587edb60b9a\") " Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.626756 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config" (OuterVolumeSpecName: "config") pod "101dce53-f7ef-4b92-9c0b-c587edb60b9a" (UID: "101dce53-f7ef-4b92-9c0b-c587edb60b9a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.626974 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "101dce53-f7ef-4b92-9c0b-c587edb60b9a" (UID: "101dce53-f7ef-4b92-9c0b-c587edb60b9a"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.627049 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "101dce53-f7ef-4b92-9c0b-c587edb60b9a" (UID: "101dce53-f7ef-4b92-9c0b-c587edb60b9a"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.627113 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101dce53-f7ef-4b92-9c0b-c587edb60b9a-datadir" (OuterVolumeSpecName: "datadir") pod "101dce53-f7ef-4b92-9c0b-c587edb60b9a" (UID: "101dce53-f7ef-4b92-9c0b-c587edb60b9a"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.627127 4767 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-entrypoint\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.627151 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.627322 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "101dce53-f7ef-4b92-9c0b-c587edb60b9a" (UID: "101dce53-f7ef-4b92-9c0b-c587edb60b9a"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.630011 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "101dce53-f7ef-4b92-9c0b-c587edb60b9a" (UID: "101dce53-f7ef-4b92-9c0b-c587edb60b9a"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.632071 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-token" (OuterVolumeSpecName: "collector-token") pod "101dce53-f7ef-4b92-9c0b-c587edb60b9a" (UID: "101dce53-f7ef-4b92-9c0b-c587edb60b9a"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.632147 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-metrics" (OuterVolumeSpecName: "metrics") pod "101dce53-f7ef-4b92-9c0b-c587edb60b9a" (UID: "101dce53-f7ef-4b92-9c0b-c587edb60b9a"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.632222 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/101dce53-f7ef-4b92-9c0b-c587edb60b9a-tmp" (OuterVolumeSpecName: "tmp") pod "101dce53-f7ef-4b92-9c0b-c587edb60b9a" (UID: "101dce53-f7ef-4b92-9c0b-c587edb60b9a"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.632545 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-sa-token" (OuterVolumeSpecName: "sa-token") pod "101dce53-f7ef-4b92-9c0b-c587edb60b9a" (UID: "101dce53-f7ef-4b92-9c0b-c587edb60b9a"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.634405 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-kube-api-access-h8cdj" (OuterVolumeSpecName: "kube-api-access-h8cdj") pod "101dce53-f7ef-4b92-9c0b-c587edb60b9a" (UID: "101dce53-f7ef-4b92-9c0b-c587edb60b9a"). InnerVolumeSpecName "kube-api-access-h8cdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.728010 4767 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/101dce53-f7ef-4b92-9c0b-c587edb60b9a-datadir\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.728068 4767 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/101dce53-f7ef-4b92-9c0b-c587edb60b9a-tmp\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.728078 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.728087 4767 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-token\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.728103 4767 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.728114 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8cdj\" (UniqueName: \"kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-kube-api-access-h8cdj\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.728122 4767 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/101dce53-f7ef-4b92-9c0b-c587edb60b9a-metrics\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.728131 4767 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/101dce53-f7ef-4b92-9c0b-c587edb60b9a-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:11 crc kubenswrapper[4767]: I0317 15:54:11.728140 4767 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/101dce53-f7ef-4b92-9c0b-c587edb60b9a-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.594323 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-gdslt" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.652013 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-gdslt"] Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.658985 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-gdslt"] Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.669902 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-4wr62"] Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.692521 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-4wr62"] Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.692720 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.697982 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-n2c2q" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.698962 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.699089 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.700111 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.700694 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.712294 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.747164 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-config-openshift-service-cacrt\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.747251 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/aa723fa9-1322-4761-b46b-0b3ad4af761f-datadir\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.747337 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-config\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.747984 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-trusted-ca\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.748238 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/aa723fa9-1322-4761-b46b-0b3ad4af761f-sa-token\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.748428 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-entrypoint\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.748584 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/aa723fa9-1322-4761-b46b-0b3ad4af761f-metrics\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.748704 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/aa723fa9-1322-4761-b46b-0b3ad4af761f-collector-syslog-receiver\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.748845 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm4xq\" (UniqueName: \"kubernetes.io/projected/aa723fa9-1322-4761-b46b-0b3ad4af761f-kube-api-access-bm4xq\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.748997 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/aa723fa9-1322-4761-b46b-0b3ad4af761f-collector-token\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.749159 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/aa723fa9-1322-4761-b46b-0b3ad4af761f-tmp\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.850853 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-entrypoint\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.850941 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/aa723fa9-1322-4761-b46b-0b3ad4af761f-metrics\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.850962 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/aa723fa9-1322-4761-b46b-0b3ad4af761f-collector-syslog-receiver\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.850983 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm4xq\" (UniqueName: \"kubernetes.io/projected/aa723fa9-1322-4761-b46b-0b3ad4af761f-kube-api-access-bm4xq\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.851014 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/aa723fa9-1322-4761-b46b-0b3ad4af761f-collector-token\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.851042 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/aa723fa9-1322-4761-b46b-0b3ad4af761f-tmp\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.851066 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-config-openshift-service-cacrt\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.851084 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/aa723fa9-1322-4761-b46b-0b3ad4af761f-datadir\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.851115 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-config\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.851153 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-trusted-ca\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.851205 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/aa723fa9-1322-4761-b46b-0b3ad4af761f-sa-token\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.851898 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/aa723fa9-1322-4761-b46b-0b3ad4af761f-datadir\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.852971 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-entrypoint\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.853002 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-config-openshift-service-cacrt\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.853253 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-trusted-ca\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.853422 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa723fa9-1322-4761-b46b-0b3ad4af761f-config\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.858212 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/aa723fa9-1322-4761-b46b-0b3ad4af761f-tmp\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.858225 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/aa723fa9-1322-4761-b46b-0b3ad4af761f-collector-syslog-receiver\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.863845 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/aa723fa9-1322-4761-b46b-0b3ad4af761f-metrics\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.863877 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/aa723fa9-1322-4761-b46b-0b3ad4af761f-collector-token\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.913291 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm4xq\" (UniqueName: \"kubernetes.io/projected/aa723fa9-1322-4761-b46b-0b3ad4af761f-kube-api-access-bm4xq\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:12 crc kubenswrapper[4767]: I0317 15:54:12.916907 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/aa723fa9-1322-4761-b46b-0b3ad4af761f-sa-token\") pod \"collector-4wr62\" (UID: \"aa723fa9-1322-4761-b46b-0b3ad4af761f\") " pod="openshift-logging/collector-4wr62" Mar 17 15:54:13 crc kubenswrapper[4767]: I0317 15:54:13.031261 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-4wr62" Mar 17 15:54:13 crc kubenswrapper[4767]: I0317 15:54:13.365495 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="101dce53-f7ef-4b92-9c0b-c587edb60b9a" path="/var/lib/kubelet/pods/101dce53-f7ef-4b92-9c0b-c587edb60b9a/volumes" Mar 17 15:54:13 crc kubenswrapper[4767]: I0317 15:54:13.492969 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-4wr62"] Mar 17 15:54:13 crc kubenswrapper[4767]: I0317 15:54:13.603873 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-4wr62" event={"ID":"aa723fa9-1322-4761-b46b-0b3ad4af761f","Type":"ContainerStarted","Data":"91b1d5daf20cbf596319bc22ffce8942f5f261fdcfa7d41253d0a41d22f92eb8"} Mar 17 15:54:18 crc kubenswrapper[4767]: I0317 15:54:18.663491 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-4wr62" event={"ID":"aa723fa9-1322-4761-b46b-0b3ad4af761f","Type":"ContainerStarted","Data":"e27be52160cbf80d903978c8d251d3b883b191b3a6edfab2a813f7dadc916577"} Mar 17 15:54:18 crc kubenswrapper[4767]: I0317 15:54:18.698649 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-4wr62" podStartSLOduration=2.590890761 podStartE2EDuration="6.698617131s" podCreationTimestamp="2026-03-17 15:54:12 +0000 UTC" firstStartedPulling="2026-03-17 15:54:13.518340123 +0000 UTC m=+1044.931656170" lastFinishedPulling="2026-03-17 15:54:17.626066493 +0000 UTC m=+1049.039382540" observedRunningTime="2026-03-17 15:54:18.692481522 +0000 UTC m=+1050.105797589" watchObservedRunningTime="2026-03-17 15:54:18.698617131 +0000 UTC m=+1050.111933178" Mar 17 15:54:32 crc kubenswrapper[4767]: I0317 15:54:32.850503 4767 scope.go:117] "RemoveContainer" containerID="be17fb93e42627d48a838082c1c40150df0a32372ee56d6c5f699c7fad43a228" Mar 17 15:54:34 crc kubenswrapper[4767]: I0317 15:54:34.167152 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:54:34 crc kubenswrapper[4767]: I0317 15:54:34.167715 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:54:34 crc kubenswrapper[4767]: I0317 15:54:34.167776 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:54:34 crc kubenswrapper[4767]: I0317 15:54:34.168438 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"265b2d9bdc4eb65af73c1e843186a34dfa697843aaf572fda7ed43575c18359a"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 15:54:34 crc kubenswrapper[4767]: I0317 15:54:34.168500 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://265b2d9bdc4eb65af73c1e843186a34dfa697843aaf572fda7ed43575c18359a" gracePeriod=600 Mar 17 15:54:34 crc kubenswrapper[4767]: I0317 15:54:34.835489 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="265b2d9bdc4eb65af73c1e843186a34dfa697843aaf572fda7ed43575c18359a" exitCode=0 Mar 17 15:54:34 crc kubenswrapper[4767]: I0317 15:54:34.836456 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"265b2d9bdc4eb65af73c1e843186a34dfa697843aaf572fda7ed43575c18359a"} Mar 17 15:54:34 crc kubenswrapper[4767]: I0317 15:54:34.836506 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"c0a59e1cd71c114bacc3d0ebc2a7a84e16351d4b20330e3bb7479cc322816c8b"} Mar 17 15:54:34 crc kubenswrapper[4767]: I0317 15:54:34.836527 4767 scope.go:117] "RemoveContainer" containerID="942ec751e9132c97b85320d70dbc413daf5abb6e5090f70bbebaf738b7e2ef34" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.157106 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh"] Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.159761 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.162675 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.171506 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh"] Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.328593 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.328677 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.328827 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbdc6\" (UniqueName: \"kubernetes.io/projected/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-kube-api-access-xbdc6\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.431130 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.431325 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.431418 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbdc6\" (UniqueName: \"kubernetes.io/projected/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-kube-api-access-xbdc6\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.432506 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.432535 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.461256 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbdc6\" (UniqueName: \"kubernetes.io/projected/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-kube-api-access-xbdc6\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.493689 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:51 crc kubenswrapper[4767]: I0317 15:54:51.980606 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh"] Mar 17 15:54:51 crc kubenswrapper[4767]: W0317 15:54:51.988420 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd16f97d_e1bf_4fe4_ac96_63468f9234b7.slice/crio-b2853730ffce688e56e8436b19f0d38640612623adf604eaa8b3e450e5f5c4d8 WatchSource:0}: Error finding container b2853730ffce688e56e8436b19f0d38640612623adf604eaa8b3e450e5f5c4d8: Status 404 returned error can't find the container with id b2853730ffce688e56e8436b19f0d38640612623adf604eaa8b3e450e5f5c4d8 Mar 17 15:54:52 crc kubenswrapper[4767]: I0317 15:54:52.021399 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" event={"ID":"cd16f97d-e1bf-4fe4-ac96-63468f9234b7","Type":"ContainerStarted","Data":"b2853730ffce688e56e8436b19f0d38640612623adf604eaa8b3e450e5f5c4d8"} Mar 17 15:54:53 crc kubenswrapper[4767]: I0317 15:54:53.030824 4767 generic.go:334] "Generic (PLEG): container finished" podID="cd16f97d-e1bf-4fe4-ac96-63468f9234b7" containerID="07c9ce001d9a9c350e63b449755f229453cfbde8873f8980a1ff9a1d265a51a2" exitCode=0 Mar 17 15:54:53 crc kubenswrapper[4767]: I0317 15:54:53.030966 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" event={"ID":"cd16f97d-e1bf-4fe4-ac96-63468f9234b7","Type":"ContainerDied","Data":"07c9ce001d9a9c350e63b449755f229453cfbde8873f8980a1ff9a1d265a51a2"} Mar 17 15:54:55 crc kubenswrapper[4767]: I0317 15:54:55.060661 4767 generic.go:334] "Generic (PLEG): container finished" podID="cd16f97d-e1bf-4fe4-ac96-63468f9234b7" containerID="c0d639039ec5af7169f1dd4bf08cef7b3c82b250671e4ce5cf662238a919a7c9" exitCode=0 Mar 17 15:54:55 crc kubenswrapper[4767]: I0317 15:54:55.060752 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" event={"ID":"cd16f97d-e1bf-4fe4-ac96-63468f9234b7","Type":"ContainerDied","Data":"c0d639039ec5af7169f1dd4bf08cef7b3c82b250671e4ce5cf662238a919a7c9"} Mar 17 15:54:56 crc kubenswrapper[4767]: I0317 15:54:56.071925 4767 generic.go:334] "Generic (PLEG): container finished" podID="cd16f97d-e1bf-4fe4-ac96-63468f9234b7" containerID="9df10922095d3074157126b72b41191d90139b9542ca8ba76d55c36eb8674973" exitCode=0 Mar 17 15:54:56 crc kubenswrapper[4767]: I0317 15:54:56.072053 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" event={"ID":"cd16f97d-e1bf-4fe4-ac96-63468f9234b7","Type":"ContainerDied","Data":"9df10922095d3074157126b72b41191d90139b9542ca8ba76d55c36eb8674973"} Mar 17 15:54:57 crc kubenswrapper[4767]: I0317 15:54:57.401687 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:54:57 crc kubenswrapper[4767]: I0317 15:54:57.587651 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-util\") pod \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " Mar 17 15:54:57 crc kubenswrapper[4767]: I0317 15:54:57.588396 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbdc6\" (UniqueName: \"kubernetes.io/projected/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-kube-api-access-xbdc6\") pod \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " Mar 17 15:54:57 crc kubenswrapper[4767]: I0317 15:54:57.588640 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-bundle\") pod \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\" (UID: \"cd16f97d-e1bf-4fe4-ac96-63468f9234b7\") " Mar 17 15:54:57 crc kubenswrapper[4767]: I0317 15:54:57.589275 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-bundle" (OuterVolumeSpecName: "bundle") pod "cd16f97d-e1bf-4fe4-ac96-63468f9234b7" (UID: "cd16f97d-e1bf-4fe4-ac96-63468f9234b7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:54:57 crc kubenswrapper[4767]: I0317 15:54:57.595478 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-kube-api-access-xbdc6" (OuterVolumeSpecName: "kube-api-access-xbdc6") pod "cd16f97d-e1bf-4fe4-ac96-63468f9234b7" (UID: "cd16f97d-e1bf-4fe4-ac96-63468f9234b7"). InnerVolumeSpecName "kube-api-access-xbdc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:54:57 crc kubenswrapper[4767]: I0317 15:54:57.599289 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-util" (OuterVolumeSpecName: "util") pod "cd16f97d-e1bf-4fe4-ac96-63468f9234b7" (UID: "cd16f97d-e1bf-4fe4-ac96-63468f9234b7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:54:57 crc kubenswrapper[4767]: I0317 15:54:57.690991 4767 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-util\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:57 crc kubenswrapper[4767]: I0317 15:54:57.691049 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbdc6\" (UniqueName: \"kubernetes.io/projected/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-kube-api-access-xbdc6\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:57 crc kubenswrapper[4767]: I0317 15:54:57.691067 4767 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd16f97d-e1bf-4fe4-ac96-63468f9234b7-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:54:58 crc kubenswrapper[4767]: I0317 15:54:58.101840 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" event={"ID":"cd16f97d-e1bf-4fe4-ac96-63468f9234b7","Type":"ContainerDied","Data":"b2853730ffce688e56e8436b19f0d38640612623adf604eaa8b3e450e5f5c4d8"} Mar 17 15:54:58 crc kubenswrapper[4767]: I0317 15:54:58.101902 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2853730ffce688e56e8436b19f0d38640612623adf604eaa8b3e450e5f5c4d8" Mar 17 15:54:58 crc kubenswrapper[4767]: I0317 15:54:58.101954 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874vq5sh" Mar 17 15:55:03 crc kubenswrapper[4767]: I0317 15:55:03.830422 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-jpjcc"] Mar 17 15:55:03 crc kubenswrapper[4767]: E0317 15:55:03.832767 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd16f97d-e1bf-4fe4-ac96-63468f9234b7" containerName="util" Mar 17 15:55:03 crc kubenswrapper[4767]: I0317 15:55:03.832866 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd16f97d-e1bf-4fe4-ac96-63468f9234b7" containerName="util" Mar 17 15:55:03 crc kubenswrapper[4767]: E0317 15:55:03.832952 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd16f97d-e1bf-4fe4-ac96-63468f9234b7" containerName="extract" Mar 17 15:55:03 crc kubenswrapper[4767]: I0317 15:55:03.833031 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd16f97d-e1bf-4fe4-ac96-63468f9234b7" containerName="extract" Mar 17 15:55:03 crc kubenswrapper[4767]: E0317 15:55:03.833120 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd16f97d-e1bf-4fe4-ac96-63468f9234b7" containerName="pull" Mar 17 15:55:03 crc kubenswrapper[4767]: I0317 15:55:03.833217 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd16f97d-e1bf-4fe4-ac96-63468f9234b7" containerName="pull" Mar 17 15:55:03 crc kubenswrapper[4767]: I0317 15:55:03.833476 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd16f97d-e1bf-4fe4-ac96-63468f9234b7" containerName="extract" Mar 17 15:55:03 crc kubenswrapper[4767]: I0317 15:55:03.835928 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jpjcc" Mar 17 15:55:03 crc kubenswrapper[4767]: I0317 15:55:03.933984 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9ct2\" (UniqueName: \"kubernetes.io/projected/c14185aa-2023-4e10-b536-1a945437b4f1-kube-api-access-f9ct2\") pod \"nmstate-operator-796d4cfff4-jpjcc\" (UID: \"c14185aa-2023-4e10-b536-1a945437b4f1\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-jpjcc" Mar 17 15:55:04 crc kubenswrapper[4767]: I0317 15:55:04.123751 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 17 15:55:04 crc kubenswrapper[4767]: I0317 15:55:04.124601 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 17 15:55:04 crc kubenswrapper[4767]: I0317 15:55:04.125420 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9ct2\" (UniqueName: \"kubernetes.io/projected/c14185aa-2023-4e10-b536-1a945437b4f1-kube-api-access-f9ct2\") pod \"nmstate-operator-796d4cfff4-jpjcc\" (UID: \"c14185aa-2023-4e10-b536-1a945437b4f1\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-jpjcc" Mar 17 15:55:04 crc kubenswrapper[4767]: I0317 15:55:04.177832 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-jpjcc"] Mar 17 15:55:04 crc kubenswrapper[4767]: I0317 15:55:04.179868 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9ct2\" (UniqueName: \"kubernetes.io/projected/c14185aa-2023-4e10-b536-1a945437b4f1-kube-api-access-f9ct2\") pod \"nmstate-operator-796d4cfff4-jpjcc\" (UID: \"c14185aa-2023-4e10-b536-1a945437b4f1\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-jpjcc" Mar 17 15:55:04 crc kubenswrapper[4767]: I0317 15:55:04.194823 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-m826c" Mar 17 15:55:04 crc kubenswrapper[4767]: I0317 15:55:04.286394 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jpjcc" Mar 17 15:55:04 crc kubenswrapper[4767]: I0317 15:55:04.957237 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-jpjcc"] Mar 17 15:55:05 crc kubenswrapper[4767]: I0317 15:55:05.306118 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jpjcc" event={"ID":"c14185aa-2023-4e10-b536-1a945437b4f1","Type":"ContainerStarted","Data":"cda4e68e952e98314f21252b6e76269c22b7f1eb933dfc44004f814fc8087dc0"} Mar 17 15:55:08 crc kubenswrapper[4767]: I0317 15:55:08.335107 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jpjcc" event={"ID":"c14185aa-2023-4e10-b536-1a945437b4f1","Type":"ContainerStarted","Data":"afe1fc1a4756f05b01dcdd90a0eb5a31386fb92f1d2576223c4dd5a184ad9c16"} Mar 17 15:55:08 crc kubenswrapper[4767]: I0317 15:55:08.363869 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jpjcc" podStartSLOduration=2.876741365 podStartE2EDuration="5.36383624s" podCreationTimestamp="2026-03-17 15:55:03 +0000 UTC" firstStartedPulling="2026-03-17 15:55:04.976112833 +0000 UTC m=+1096.389428880" lastFinishedPulling="2026-03-17 15:55:07.463207718 +0000 UTC m=+1098.876523755" observedRunningTime="2026-03-17 15:55:08.356145916 +0000 UTC m=+1099.769461973" watchObservedRunningTime="2026-03-17 15:55:08.36383624 +0000 UTC m=+1099.777152287" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.472880 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w"] Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.474956 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.482884 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-gn6vg" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.501075 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.505636 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n4r6\" (UniqueName: \"kubernetes.io/projected/31f30608-3b0d-4f63-9ab8-6a1547e233ff-kube-api-access-8n4r6\") pod \"nmstate-webhook-5f558f5558-rwq8w\" (UID: \"31f30608-3b0d-4f63-9ab8-6a1547e233ff\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.505763 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/31f30608-3b0d-4f63-9ab8-6a1547e233ff-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-rwq8w\" (UID: \"31f30608-3b0d-4f63-9ab8-6a1547e233ff\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.508674 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w"] Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.555369 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m"] Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.558233 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.669904 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/31f30608-3b0d-4f63-9ab8-6a1547e233ff-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-rwq8w\" (UID: \"31f30608-3b0d-4f63-9ab8-6a1547e233ff\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.670415 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtkbn\" (UniqueName: \"kubernetes.io/projected/7eba85a7-b1ee-4883-a03c-c099886f2a68-kube-api-access-xtkbn\") pod \"nmstate-metrics-9b8c8685d-qnt4m\" (UID: \"7eba85a7-b1ee-4883-a03c-c099886f2a68\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.670622 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n4r6\" (UniqueName: \"kubernetes.io/projected/31f30608-3b0d-4f63-9ab8-6a1547e233ff-kube-api-access-8n4r6\") pod \"nmstate-webhook-5f558f5558-rwq8w\" (UID: \"31f30608-3b0d-4f63-9ab8-6a1547e233ff\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.671228 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-6jnww"] Mar 17 15:55:12 crc kubenswrapper[4767]: E0317 15:55:12.671299 4767 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Mar 17 15:55:12 crc kubenswrapper[4767]: E0317 15:55:12.671404 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31f30608-3b0d-4f63-9ab8-6a1547e233ff-tls-key-pair podName:31f30608-3b0d-4f63-9ab8-6a1547e233ff nodeName:}" failed. No retries permitted until 2026-03-17 15:55:13.171364197 +0000 UTC m=+1104.584680314 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/31f30608-3b0d-4f63-9ab8-6a1547e233ff-tls-key-pair") pod "nmstate-webhook-5f558f5558-rwq8w" (UID: "31f30608-3b0d-4f63-9ab8-6a1547e233ff") : secret "openshift-nmstate-webhook" not found Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.672522 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.688299 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m"] Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.728629 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n4r6\" (UniqueName: \"kubernetes.io/projected/31f30608-3b0d-4f63-9ab8-6a1547e233ff-kube-api-access-8n4r6\") pod \"nmstate-webhook-5f558f5558-rwq8w\" (UID: \"31f30608-3b0d-4f63-9ab8-6a1547e233ff\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.772647 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvsbl\" (UniqueName: \"kubernetes.io/projected/d4418b02-160e-4b8a-8232-e194eb5ac0f3-kube-api-access-zvsbl\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.772796 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtkbn\" (UniqueName: \"kubernetes.io/projected/7eba85a7-b1ee-4883-a03c-c099886f2a68-kube-api-access-xtkbn\") pod \"nmstate-metrics-9b8c8685d-qnt4m\" (UID: \"7eba85a7-b1ee-4883-a03c-c099886f2a68\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.772836 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d4418b02-160e-4b8a-8232-e194eb5ac0f3-nmstate-lock\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.772886 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d4418b02-160e-4b8a-8232-e194eb5ac0f3-dbus-socket\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.772919 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d4418b02-160e-4b8a-8232-e194eb5ac0f3-ovs-socket\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.776413 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf"] Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.777979 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.780872 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-t4gmb" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.781140 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.781846 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.811012 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf"] Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.836712 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtkbn\" (UniqueName: \"kubernetes.io/projected/7eba85a7-b1ee-4883-a03c-c099886f2a68-kube-api-access-xtkbn\") pod \"nmstate-metrics-9b8c8685d-qnt4m\" (UID: \"7eba85a7-b1ee-4883-a03c-c099886f2a68\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.874469 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/540673fb-a230-4614-a4fe-f24e3188e212-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-n7fwf\" (UID: \"540673fb-a230-4614-a4fe-f24e3188e212\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.874895 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvsbl\" (UniqueName: \"kubernetes.io/projected/d4418b02-160e-4b8a-8232-e194eb5ac0f3-kube-api-access-zvsbl\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.875038 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d4418b02-160e-4b8a-8232-e194eb5ac0f3-nmstate-lock\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.875149 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8kjr\" (UniqueName: \"kubernetes.io/projected/540673fb-a230-4614-a4fe-f24e3188e212-kube-api-access-v8kjr\") pod \"nmstate-console-plugin-86f58fcf4-n7fwf\" (UID: \"540673fb-a230-4614-a4fe-f24e3188e212\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.875345 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d4418b02-160e-4b8a-8232-e194eb5ac0f3-nmstate-lock\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.875471 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d4418b02-160e-4b8a-8232-e194eb5ac0f3-dbus-socket\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.875601 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d4418b02-160e-4b8a-8232-e194eb5ac0f3-ovs-socket\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.875782 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/540673fb-a230-4614-a4fe-f24e3188e212-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-n7fwf\" (UID: \"540673fb-a230-4614-a4fe-f24e3188e212\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.876295 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d4418b02-160e-4b8a-8232-e194eb5ac0f3-dbus-socket\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.876387 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d4418b02-160e-4b8a-8232-e194eb5ac0f3-ovs-socket\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.906731 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvsbl\" (UniqueName: \"kubernetes.io/projected/d4418b02-160e-4b8a-8232-e194eb5ac0f3-kube-api-access-zvsbl\") pod \"nmstate-handler-6jnww\" (UID: \"d4418b02-160e-4b8a-8232-e194eb5ac0f3\") " pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.972806 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5c7b759f69-p6dvt"] Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.973693 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.975635 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.982601 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c7b759f69-p6dvt"] Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.984578 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/540673fb-a230-4614-a4fe-f24e3188e212-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-n7fwf\" (UID: \"540673fb-a230-4614-a4fe-f24e3188e212\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.984709 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/540673fb-a230-4614-a4fe-f24e3188e212-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-n7fwf\" (UID: \"540673fb-a230-4614-a4fe-f24e3188e212\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.984804 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8kjr\" (UniqueName: \"kubernetes.io/projected/540673fb-a230-4614-a4fe-f24e3188e212-kube-api-access-v8kjr\") pod \"nmstate-console-plugin-86f58fcf4-n7fwf\" (UID: \"540673fb-a230-4614-a4fe-f24e3188e212\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:12 crc kubenswrapper[4767]: E0317 15:55:12.985226 4767 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 17 15:55:12 crc kubenswrapper[4767]: E0317 15:55:12.985289 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/540673fb-a230-4614-a4fe-f24e3188e212-plugin-serving-cert podName:540673fb-a230-4614-a4fe-f24e3188e212 nodeName:}" failed. No retries permitted until 2026-03-17 15:55:13.485270907 +0000 UTC m=+1104.898586954 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/540673fb-a230-4614-a4fe-f24e3188e212-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-n7fwf" (UID: "540673fb-a230-4614-a4fe-f24e3188e212") : secret "plugin-serving-cert" not found Mar 17 15:55:12 crc kubenswrapper[4767]: I0317 15:55:12.986340 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/540673fb-a230-4614-a4fe-f24e3188e212-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-n7fwf\" (UID: \"540673fb-a230-4614-a4fe-f24e3188e212\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.010805 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.034886 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8kjr\" (UniqueName: \"kubernetes.io/projected/540673fb-a230-4614-a4fe-f24e3188e212-kube-api-access-v8kjr\") pod \"nmstate-console-plugin-86f58fcf4-n7fwf\" (UID: \"540673fb-a230-4614-a4fe-f24e3188e212\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.087076 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-oauth-config\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.087743 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-oauth-serving-cert\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.087828 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-trusted-ca-bundle\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.087873 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-serving-cert\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.087928 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzln9\" (UniqueName: \"kubernetes.io/projected/63b9ae01-2768-463b-8aab-569c360824a8-kube-api-access-dzln9\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.087955 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-service-ca\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.088059 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-console-config\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.194937 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-trusted-ca-bundle\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.196132 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-trusted-ca-bundle\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.196221 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-serving-cert\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.196382 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzln9\" (UniqueName: \"kubernetes.io/projected/63b9ae01-2768-463b-8aab-569c360824a8-kube-api-access-dzln9\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.196408 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-service-ca\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.196501 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-console-config\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.196623 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/31f30608-3b0d-4f63-9ab8-6a1547e233ff-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-rwq8w\" (UID: \"31f30608-3b0d-4f63-9ab8-6a1547e233ff\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.196707 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-oauth-config\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.196785 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-oauth-serving-cert\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.198069 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-oauth-serving-cert\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.200070 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-console-config\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.200804 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-service-ca\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.206256 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-serving-cert\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.212683 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-oauth-config\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.228153 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/31f30608-3b0d-4f63-9ab8-6a1547e233ff-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-rwq8w\" (UID: \"31f30608-3b0d-4f63-9ab8-6a1547e233ff\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.229974 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzln9\" (UniqueName: \"kubernetes.io/projected/63b9ae01-2768-463b-8aab-569c360824a8-kube-api-access-dzln9\") pod \"console-5c7b759f69-p6dvt\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.298725 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.399872 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.424019 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6jnww" event={"ID":"d4418b02-160e-4b8a-8232-e194eb5ac0f3","Type":"ContainerStarted","Data":"6a2563b1ae908de8021ee5b4ada553da21f4e49ca83dde26f15c3c6c84191fc2"} Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.653877 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/540673fb-a230-4614-a4fe-f24e3188e212-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-n7fwf\" (UID: \"540673fb-a230-4614-a4fe-f24e3188e212\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.661056 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/540673fb-a230-4614-a4fe-f24e3188e212-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-n7fwf\" (UID: \"540673fb-a230-4614-a4fe-f24e3188e212\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.706725 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" Mar 17 15:55:13 crc kubenswrapper[4767]: I0317 15:55:13.724301 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m"] Mar 17 15:55:13 crc kubenswrapper[4767]: W0317 15:55:13.745086 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7eba85a7_b1ee_4883_a03c_c099886f2a68.slice/crio-a84694dc294245fadda04121687d16caed7db7f1125599e5c33a0979385a8d6c WatchSource:0}: Error finding container a84694dc294245fadda04121687d16caed7db7f1125599e5c33a0979385a8d6c: Status 404 returned error can't find the container with id a84694dc294245fadda04121687d16caed7db7f1125599e5c33a0979385a8d6c Mar 17 15:55:14 crc kubenswrapper[4767]: I0317 15:55:14.172670 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf"] Mar 17 15:55:14 crc kubenswrapper[4767]: W0317 15:55:14.189135 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod540673fb_a230_4614_a4fe_f24e3188e212.slice/crio-f28fe22dddea4008f961fdf5a7a3c55e1136d87df481505d5f45204456f0caaf WatchSource:0}: Error finding container f28fe22dddea4008f961fdf5a7a3c55e1136d87df481505d5f45204456f0caaf: Status 404 returned error can't find the container with id f28fe22dddea4008f961fdf5a7a3c55e1136d87df481505d5f45204456f0caaf Mar 17 15:55:14 crc kubenswrapper[4767]: I0317 15:55:14.278443 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c7b759f69-p6dvt"] Mar 17 15:55:14 crc kubenswrapper[4767]: W0317 15:55:14.287405 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63b9ae01_2768_463b_8aab_569c360824a8.slice/crio-a003cd890c236aedefd25287e0942d73fe5daab8653ca1eaf738a15887ce6c7a WatchSource:0}: Error finding container a003cd890c236aedefd25287e0942d73fe5daab8653ca1eaf738a15887ce6c7a: Status 404 returned error can't find the container with id a003cd890c236aedefd25287e0942d73fe5daab8653ca1eaf738a15887ce6c7a Mar 17 15:55:14 crc kubenswrapper[4767]: I0317 15:55:14.300671 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w"] Mar 17 15:55:14 crc kubenswrapper[4767]: W0317 15:55:14.304773 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31f30608_3b0d_4f63_9ab8_6a1547e233ff.slice/crio-76ae2edc6b6a555c8f51314e15f21c7538a2ea0c1977ebec29e630cff11f446e WatchSource:0}: Error finding container 76ae2edc6b6a555c8f51314e15f21c7538a2ea0c1977ebec29e630cff11f446e: Status 404 returned error can't find the container with id 76ae2edc6b6a555c8f51314e15f21c7538a2ea0c1977ebec29e630cff11f446e Mar 17 15:55:14 crc kubenswrapper[4767]: I0317 15:55:14.432239 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c7b759f69-p6dvt" event={"ID":"63b9ae01-2768-463b-8aab-569c360824a8","Type":"ContainerStarted","Data":"a003cd890c236aedefd25287e0942d73fe5daab8653ca1eaf738a15887ce6c7a"} Mar 17 15:55:14 crc kubenswrapper[4767]: I0317 15:55:14.435720 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" event={"ID":"540673fb-a230-4614-a4fe-f24e3188e212","Type":"ContainerStarted","Data":"f28fe22dddea4008f961fdf5a7a3c55e1136d87df481505d5f45204456f0caaf"} Mar 17 15:55:14 crc kubenswrapper[4767]: I0317 15:55:14.436792 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" event={"ID":"31f30608-3b0d-4f63-9ab8-6a1547e233ff","Type":"ContainerStarted","Data":"76ae2edc6b6a555c8f51314e15f21c7538a2ea0c1977ebec29e630cff11f446e"} Mar 17 15:55:14 crc kubenswrapper[4767]: I0317 15:55:14.438227 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m" event={"ID":"7eba85a7-b1ee-4883-a03c-c099886f2a68","Type":"ContainerStarted","Data":"a84694dc294245fadda04121687d16caed7db7f1125599e5c33a0979385a8d6c"} Mar 17 15:55:15 crc kubenswrapper[4767]: I0317 15:55:15.456360 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c7b759f69-p6dvt" event={"ID":"63b9ae01-2768-463b-8aab-569c360824a8","Type":"ContainerStarted","Data":"2273e1025467dcb4d2442bd5702080c625fd5a89a6060769921396296113365e"} Mar 17 15:55:15 crc kubenswrapper[4767]: I0317 15:55:15.485101 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5c7b759f69-p6dvt" podStartSLOduration=3.485071391 podStartE2EDuration="3.485071391s" podCreationTimestamp="2026-03-17 15:55:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:55:15.482459075 +0000 UTC m=+1106.895775132" watchObservedRunningTime="2026-03-17 15:55:15.485071391 +0000 UTC m=+1106.898387458" Mar 17 15:55:18 crc kubenswrapper[4767]: I0317 15:55:18.490211 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" event={"ID":"540673fb-a230-4614-a4fe-f24e3188e212","Type":"ContainerStarted","Data":"9e28b7847b440877cdbdad2c23fcef07d6c202df27bb517241a82c5954318027"} Mar 17 15:55:18 crc kubenswrapper[4767]: I0317 15:55:18.493550 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" event={"ID":"31f30608-3b0d-4f63-9ab8-6a1547e233ff","Type":"ContainerStarted","Data":"7987c2d6d8ce6e2bf0402c0034545db4f760f5b265146e9d2b8aafdd99922ea0"} Mar 17 15:55:18 crc kubenswrapper[4767]: I0317 15:55:18.493706 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 15:55:18 crc kubenswrapper[4767]: I0317 15:55:18.497036 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6jnww" event={"ID":"d4418b02-160e-4b8a-8232-e194eb5ac0f3","Type":"ContainerStarted","Data":"0314eb8e3617e2e728cdb9fb32d18693b42a21478eb680d2607b87e2120f8a9b"} Mar 17 15:55:18 crc kubenswrapper[4767]: I0317 15:55:18.497186 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:18 crc kubenswrapper[4767]: I0317 15:55:18.499965 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m" event={"ID":"7eba85a7-b1ee-4883-a03c-c099886f2a68","Type":"ContainerStarted","Data":"0aa015d87a945d34cf90f2491a2838c4b741f663d254228e3299fbe9285b9a2f"} Mar 17 15:55:18 crc kubenswrapper[4767]: I0317 15:55:18.510762 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-n7fwf" podStartSLOduration=2.634830946 podStartE2EDuration="6.510732856s" podCreationTimestamp="2026-03-17 15:55:12 +0000 UTC" firstStartedPulling="2026-03-17 15:55:14.193489976 +0000 UTC m=+1105.606806023" lastFinishedPulling="2026-03-17 15:55:18.069391886 +0000 UTC m=+1109.482707933" observedRunningTime="2026-03-17 15:55:18.50776813 +0000 UTC m=+1109.921084177" watchObservedRunningTime="2026-03-17 15:55:18.510732856 +0000 UTC m=+1109.924048923" Mar 17 15:55:18 crc kubenswrapper[4767]: I0317 15:55:18.539407 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" podStartSLOduration=2.784784282 podStartE2EDuration="6.53937208s" podCreationTimestamp="2026-03-17 15:55:12 +0000 UTC" firstStartedPulling="2026-03-17 15:55:14.313394397 +0000 UTC m=+1105.726710434" lastFinishedPulling="2026-03-17 15:55:18.067982185 +0000 UTC m=+1109.481298232" observedRunningTime="2026-03-17 15:55:18.532235882 +0000 UTC m=+1109.945551949" watchObservedRunningTime="2026-03-17 15:55:18.53937208 +0000 UTC m=+1109.952688127" Mar 17 15:55:18 crc kubenswrapper[4767]: I0317 15:55:18.565931 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-6jnww" podStartSLOduration=1.631441911 podStartE2EDuration="6.565891472s" podCreationTimestamp="2026-03-17 15:55:12 +0000 UTC" firstStartedPulling="2026-03-17 15:55:13.147415568 +0000 UTC m=+1104.560731615" lastFinishedPulling="2026-03-17 15:55:18.081865129 +0000 UTC m=+1109.495181176" observedRunningTime="2026-03-17 15:55:18.554876622 +0000 UTC m=+1109.968192679" watchObservedRunningTime="2026-03-17 15:55:18.565891472 +0000 UTC m=+1109.979207539" Mar 17 15:55:21 crc kubenswrapper[4767]: I0317 15:55:21.634575 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m" event={"ID":"7eba85a7-b1ee-4883-a03c-c099886f2a68","Type":"ContainerStarted","Data":"6baefa31041d047be2f9298b8df42384cf6dbe7a505ce4329940ea6151669f80"} Mar 17 15:55:23 crc kubenswrapper[4767]: I0317 15:55:23.042458 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-6jnww" Mar 17 15:55:23 crc kubenswrapper[4767]: I0317 15:55:23.067079 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-qnt4m" podStartSLOduration=3.702614145 podStartE2EDuration="11.067045467s" podCreationTimestamp="2026-03-17 15:55:12 +0000 UTC" firstStartedPulling="2026-03-17 15:55:13.773483487 +0000 UTC m=+1105.186799544" lastFinishedPulling="2026-03-17 15:55:21.137914819 +0000 UTC m=+1112.551230866" observedRunningTime="2026-03-17 15:55:21.680538088 +0000 UTC m=+1113.093854155" watchObservedRunningTime="2026-03-17 15:55:23.067045467 +0000 UTC m=+1114.480361524" Mar 17 15:55:23 crc kubenswrapper[4767]: I0317 15:55:23.300260 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:23 crc kubenswrapper[4767]: I0317 15:55:23.300856 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:23 crc kubenswrapper[4767]: I0317 15:55:23.306564 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:23 crc kubenswrapper[4767]: I0317 15:55:23.656214 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 15:55:23 crc kubenswrapper[4767]: I0317 15:55:23.843722 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-55c858d47f-r899j"] Mar 17 15:55:33 crc kubenswrapper[4767]: I0317 15:55:33.406345 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 15:55:48 crc kubenswrapper[4767]: I0317 15:55:48.936534 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-55c858d47f-r899j" podUID="ea4340bd-c8ea-4286-be77-f537b9f0d99d" containerName="console" containerID="cri-o://ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b" gracePeriod=15 Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.550661 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-55c858d47f-r899j_ea4340bd-c8ea-4286-be77-f537b9f0d99d/console/0.log" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.551638 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.730098 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-trusted-ca-bundle\") pod \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.730216 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-oauth-config\") pod \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.730244 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-service-ca\") pod \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.730279 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-config\") pod \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.730381 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pkj7\" (UniqueName: \"kubernetes.io/projected/ea4340bd-c8ea-4286-be77-f537b9f0d99d-kube-api-access-2pkj7\") pod \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.730417 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-oauth-serving-cert\") pod \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.730480 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-serving-cert\") pod \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\" (UID: \"ea4340bd-c8ea-4286-be77-f537b9f0d99d\") " Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.731312 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ea4340bd-c8ea-4286-be77-f537b9f0d99d" (UID: "ea4340bd-c8ea-4286-be77-f537b9f0d99d"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.731367 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-service-ca" (OuterVolumeSpecName: "service-ca") pod "ea4340bd-c8ea-4286-be77-f537b9f0d99d" (UID: "ea4340bd-c8ea-4286-be77-f537b9f0d99d"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.731759 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-config" (OuterVolumeSpecName: "console-config") pod "ea4340bd-c8ea-4286-be77-f537b9f0d99d" (UID: "ea4340bd-c8ea-4286-be77-f537b9f0d99d"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.732008 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ea4340bd-c8ea-4286-be77-f537b9f0d99d" (UID: "ea4340bd-c8ea-4286-be77-f537b9f0d99d"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.737750 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ea4340bd-c8ea-4286-be77-f537b9f0d99d" (UID: "ea4340bd-c8ea-4286-be77-f537b9f0d99d"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.739773 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea4340bd-c8ea-4286-be77-f537b9f0d99d-kube-api-access-2pkj7" (OuterVolumeSpecName: "kube-api-access-2pkj7") pod "ea4340bd-c8ea-4286-be77-f537b9f0d99d" (UID: "ea4340bd-c8ea-4286-be77-f537b9f0d99d"). InnerVolumeSpecName "kube-api-access-2pkj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.747284 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ea4340bd-c8ea-4286-be77-f537b9f0d99d" (UID: "ea4340bd-c8ea-4286-be77-f537b9f0d99d"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.832456 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pkj7\" (UniqueName: \"kubernetes.io/projected/ea4340bd-c8ea-4286-be77-f537b9f0d99d-kube-api-access-2pkj7\") on node \"crc\" DevicePath \"\"" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.833096 4767 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.833114 4767 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.833144 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.833157 4767 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.833195 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.833211 4767 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ea4340bd-c8ea-4286-be77-f537b9f0d99d-console-config\") on node \"crc\" DevicePath \"\"" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.955578 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-55c858d47f-r899j_ea4340bd-c8ea-4286-be77-f537b9f0d99d/console/0.log" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.955642 4767 generic.go:334] "Generic (PLEG): container finished" podID="ea4340bd-c8ea-4286-be77-f537b9f0d99d" containerID="ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b" exitCode=2 Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.955687 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-55c858d47f-r899j" event={"ID":"ea4340bd-c8ea-4286-be77-f537b9f0d99d","Type":"ContainerDied","Data":"ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b"} Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.955735 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-55c858d47f-r899j" event={"ID":"ea4340bd-c8ea-4286-be77-f537b9f0d99d","Type":"ContainerDied","Data":"dc6c53fb145841b1b049d8ca462c2a9f40767678638a549527e55092d48df97b"} Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.955777 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-55c858d47f-r899j" Mar 17 15:55:49 crc kubenswrapper[4767]: I0317 15:55:49.955775 4767 scope.go:117] "RemoveContainer" containerID="ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b" Mar 17 15:55:50 crc kubenswrapper[4767]: I0317 15:55:50.001303 4767 scope.go:117] "RemoveContainer" containerID="ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b" Mar 17 15:55:50 crc kubenswrapper[4767]: E0317 15:55:50.002044 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b\": container with ID starting with ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b not found: ID does not exist" containerID="ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b" Mar 17 15:55:50 crc kubenswrapper[4767]: I0317 15:55:50.002140 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b"} err="failed to get container status \"ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b\": rpc error: code = NotFound desc = could not find container \"ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b\": container with ID starting with ea36b13ec51778f90022e70370a7e9264abf594486b7d0cfe291bdebd3f2de4b not found: ID does not exist" Mar 17 15:55:50 crc kubenswrapper[4767]: I0317 15:55:50.024001 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-55c858d47f-r899j"] Mar 17 15:55:50 crc kubenswrapper[4767]: I0317 15:55:50.041626 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-55c858d47f-r899j"] Mar 17 15:55:51 crc kubenswrapper[4767]: I0317 15:55:51.372707 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea4340bd-c8ea-4286-be77-f537b9f0d99d" path="/var/lib/kubelet/pods/ea4340bd-c8ea-4286-be77-f537b9f0d99d/volumes" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.683406 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx"] Mar 17 15:55:52 crc kubenswrapper[4767]: E0317 15:55:52.684360 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea4340bd-c8ea-4286-be77-f537b9f0d99d" containerName="console" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.684382 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea4340bd-c8ea-4286-be77-f537b9f0d99d" containerName="console" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.684603 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea4340bd-c8ea-4286-be77-f537b9f0d99d" containerName="console" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.685892 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.688855 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.786025 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx"] Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.854395 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.854516 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf4vh\" (UniqueName: \"kubernetes.io/projected/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-kube-api-access-bf4vh\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.854609 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.956819 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf4vh\" (UniqueName: \"kubernetes.io/projected/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-kube-api-access-bf4vh\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.956931 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.957024 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.957683 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.957705 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:52 crc kubenswrapper[4767]: I0317 15:55:52.983550 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf4vh\" (UniqueName: \"kubernetes.io/projected/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-kube-api-access-bf4vh\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:53 crc kubenswrapper[4767]: I0317 15:55:53.006393 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:53 crc kubenswrapper[4767]: I0317 15:55:53.680087 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx"] Mar 17 15:55:53 crc kubenswrapper[4767]: I0317 15:55:53.993640 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" event={"ID":"08524a1f-5c83-4a14-8c23-9b45c88a7cf6","Type":"ContainerStarted","Data":"fb184655ae9a3796858eb6a232065eb2fab1e894990e2c3a8c427b8776f08c12"} Mar 17 15:55:53 crc kubenswrapper[4767]: I0317 15:55:53.994215 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" event={"ID":"08524a1f-5c83-4a14-8c23-9b45c88a7cf6","Type":"ContainerStarted","Data":"c8321b75b77018b2a0a9e0f36eefe9c84213c99d8454118a01e1c097cb2d3bea"} Mar 17 15:55:55 crc kubenswrapper[4767]: I0317 15:55:55.003958 4767 generic.go:334] "Generic (PLEG): container finished" podID="08524a1f-5c83-4a14-8c23-9b45c88a7cf6" containerID="fb184655ae9a3796858eb6a232065eb2fab1e894990e2c3a8c427b8776f08c12" exitCode=0 Mar 17 15:55:55 crc kubenswrapper[4767]: I0317 15:55:55.004032 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" event={"ID":"08524a1f-5c83-4a14-8c23-9b45c88a7cf6","Type":"ContainerDied","Data":"fb184655ae9a3796858eb6a232065eb2fab1e894990e2c3a8c427b8776f08c12"} Mar 17 15:55:57 crc kubenswrapper[4767]: I0317 15:55:57.021187 4767 generic.go:334] "Generic (PLEG): container finished" podID="08524a1f-5c83-4a14-8c23-9b45c88a7cf6" containerID="9233a57b0f3d9c6f73a95d62b23f7322e99fdd8447787fa5b212e3d9c88dac4b" exitCode=0 Mar 17 15:55:57 crc kubenswrapper[4767]: I0317 15:55:57.021322 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" event={"ID":"08524a1f-5c83-4a14-8c23-9b45c88a7cf6","Type":"ContainerDied","Data":"9233a57b0f3d9c6f73a95d62b23f7322e99fdd8447787fa5b212e3d9c88dac4b"} Mar 17 15:55:58 crc kubenswrapper[4767]: I0317 15:55:58.035417 4767 generic.go:334] "Generic (PLEG): container finished" podID="08524a1f-5c83-4a14-8c23-9b45c88a7cf6" containerID="1b3a688fbd3776cec9a12d8134ec66e4d8bd0a185a59289b7a2e9dfa82323a08" exitCode=0 Mar 17 15:55:58 crc kubenswrapper[4767]: I0317 15:55:58.035525 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" event={"ID":"08524a1f-5c83-4a14-8c23-9b45c88a7cf6","Type":"ContainerDied","Data":"1b3a688fbd3776cec9a12d8134ec66e4d8bd0a185a59289b7a2e9dfa82323a08"} Mar 17 15:55:59 crc kubenswrapper[4767]: I0317 15:55:59.365490 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:55:59 crc kubenswrapper[4767]: I0317 15:55:59.485866 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf4vh\" (UniqueName: \"kubernetes.io/projected/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-kube-api-access-bf4vh\") pod \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " Mar 17 15:55:59 crc kubenswrapper[4767]: I0317 15:55:59.485988 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-bundle\") pod \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " Mar 17 15:55:59 crc kubenswrapper[4767]: I0317 15:55:59.486022 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-util\") pod \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\" (UID: \"08524a1f-5c83-4a14-8c23-9b45c88a7cf6\") " Mar 17 15:55:59 crc kubenswrapper[4767]: I0317 15:55:59.487301 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-bundle" (OuterVolumeSpecName: "bundle") pod "08524a1f-5c83-4a14-8c23-9b45c88a7cf6" (UID: "08524a1f-5c83-4a14-8c23-9b45c88a7cf6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:55:59 crc kubenswrapper[4767]: I0317 15:55:59.498356 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-kube-api-access-bf4vh" (OuterVolumeSpecName: "kube-api-access-bf4vh") pod "08524a1f-5c83-4a14-8c23-9b45c88a7cf6" (UID: "08524a1f-5c83-4a14-8c23-9b45c88a7cf6"). InnerVolumeSpecName "kube-api-access-bf4vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:55:59 crc kubenswrapper[4767]: I0317 15:55:59.514128 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-util" (OuterVolumeSpecName: "util") pod "08524a1f-5c83-4a14-8c23-9b45c88a7cf6" (UID: "08524a1f-5c83-4a14-8c23-9b45c88a7cf6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:55:59 crc kubenswrapper[4767]: I0317 15:55:59.588586 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf4vh\" (UniqueName: \"kubernetes.io/projected/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-kube-api-access-bf4vh\") on node \"crc\" DevicePath \"\"" Mar 17 15:55:59 crc kubenswrapper[4767]: I0317 15:55:59.588646 4767 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:55:59 crc kubenswrapper[4767]: I0317 15:55:59.588657 4767 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08524a1f-5c83-4a14-8c23-9b45c88a7cf6-util\") on node \"crc\" DevicePath \"\"" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.056706 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" event={"ID":"08524a1f-5c83-4a14-8c23-9b45c88a7cf6","Type":"ContainerDied","Data":"c8321b75b77018b2a0a9e0f36eefe9c84213c99d8454118a01e1c097cb2d3bea"} Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.056775 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8321b75b77018b2a0a9e0f36eefe9c84213c99d8454118a01e1c097cb2d3bea" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.056785 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c16r6nx" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.141263 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562716-54ssn"] Mar 17 15:56:00 crc kubenswrapper[4767]: E0317 15:56:00.141905 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08524a1f-5c83-4a14-8c23-9b45c88a7cf6" containerName="util" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.141931 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="08524a1f-5c83-4a14-8c23-9b45c88a7cf6" containerName="util" Mar 17 15:56:00 crc kubenswrapper[4767]: E0317 15:56:00.141957 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08524a1f-5c83-4a14-8c23-9b45c88a7cf6" containerName="extract" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.141968 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="08524a1f-5c83-4a14-8c23-9b45c88a7cf6" containerName="extract" Mar 17 15:56:00 crc kubenswrapper[4767]: E0317 15:56:00.142027 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08524a1f-5c83-4a14-8c23-9b45c88a7cf6" containerName="pull" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.142037 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="08524a1f-5c83-4a14-8c23-9b45c88a7cf6" containerName="pull" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.142218 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="08524a1f-5c83-4a14-8c23-9b45c88a7cf6" containerName="extract" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.142918 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562716-54ssn" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.154368 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562716-54ssn"] Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.315007 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.315419 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.316013 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcjxq\" (UniqueName: \"kubernetes.io/projected/9d7c1cce-3729-4157-9f4e-275c46c6e436-kube-api-access-rcjxq\") pod \"auto-csr-approver-29562716-54ssn\" (UID: \"9d7c1cce-3729-4157-9f4e-275c46c6e436\") " pod="openshift-infra/auto-csr-approver-29562716-54ssn" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.316245 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.418335 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcjxq\" (UniqueName: \"kubernetes.io/projected/9d7c1cce-3729-4157-9f4e-275c46c6e436-kube-api-access-rcjxq\") pod \"auto-csr-approver-29562716-54ssn\" (UID: \"9d7c1cce-3729-4157-9f4e-275c46c6e436\") " pod="openshift-infra/auto-csr-approver-29562716-54ssn" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.443108 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcjxq\" (UniqueName: \"kubernetes.io/projected/9d7c1cce-3729-4157-9f4e-275c46c6e436-kube-api-access-rcjxq\") pod \"auto-csr-approver-29562716-54ssn\" (UID: \"9d7c1cce-3729-4157-9f4e-275c46c6e436\") " pod="openshift-infra/auto-csr-approver-29562716-54ssn" Mar 17 15:56:00 crc kubenswrapper[4767]: I0317 15:56:00.623533 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562716-54ssn" Mar 17 15:56:01 crc kubenswrapper[4767]: I0317 15:56:01.430978 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562716-54ssn"] Mar 17 15:56:02 crc kubenswrapper[4767]: I0317 15:56:02.075042 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562716-54ssn" event={"ID":"9d7c1cce-3729-4157-9f4e-275c46c6e436","Type":"ContainerStarted","Data":"90456c7cbabec9b1d010036ecdcc2dd909399d193e7b7804743ef4bec3c6c8ce"} Mar 17 15:56:03 crc kubenswrapper[4767]: I0317 15:56:03.115504 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562716-54ssn" event={"ID":"9d7c1cce-3729-4157-9f4e-275c46c6e436","Type":"ContainerStarted","Data":"d15a14c592e76d72a87f044c7fa3daad5448e5a593f3caa5fd73086a616b8c52"} Mar 17 15:56:03 crc kubenswrapper[4767]: I0317 15:56:03.144161 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562716-54ssn" podStartSLOduration=2.031960725 podStartE2EDuration="3.144138366s" podCreationTimestamp="2026-03-17 15:56:00 +0000 UTC" firstStartedPulling="2026-03-17 15:56:01.439811684 +0000 UTC m=+1152.853127741" lastFinishedPulling="2026-03-17 15:56:02.551989335 +0000 UTC m=+1153.965305382" observedRunningTime="2026-03-17 15:56:03.142536009 +0000 UTC m=+1154.555852056" watchObservedRunningTime="2026-03-17 15:56:03.144138366 +0000 UTC m=+1154.557454413" Mar 17 15:56:04 crc kubenswrapper[4767]: I0317 15:56:04.127629 4767 generic.go:334] "Generic (PLEG): container finished" podID="9d7c1cce-3729-4157-9f4e-275c46c6e436" containerID="d15a14c592e76d72a87f044c7fa3daad5448e5a593f3caa5fd73086a616b8c52" exitCode=0 Mar 17 15:56:04 crc kubenswrapper[4767]: I0317 15:56:04.127732 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562716-54ssn" event={"ID":"9d7c1cce-3729-4157-9f4e-275c46c6e436","Type":"ContainerDied","Data":"d15a14c592e76d72a87f044c7fa3daad5448e5a593f3caa5fd73086a616b8c52"} Mar 17 15:56:05 crc kubenswrapper[4767]: I0317 15:56:05.509791 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562716-54ssn" Mar 17 15:56:05 crc kubenswrapper[4767]: I0317 15:56:05.666492 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcjxq\" (UniqueName: \"kubernetes.io/projected/9d7c1cce-3729-4157-9f4e-275c46c6e436-kube-api-access-rcjxq\") pod \"9d7c1cce-3729-4157-9f4e-275c46c6e436\" (UID: \"9d7c1cce-3729-4157-9f4e-275c46c6e436\") " Mar 17 15:56:05 crc kubenswrapper[4767]: I0317 15:56:05.682664 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d7c1cce-3729-4157-9f4e-275c46c6e436-kube-api-access-rcjxq" (OuterVolumeSpecName: "kube-api-access-rcjxq") pod "9d7c1cce-3729-4157-9f4e-275c46c6e436" (UID: "9d7c1cce-3729-4157-9f4e-275c46c6e436"). InnerVolumeSpecName "kube-api-access-rcjxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:56:05 crc kubenswrapper[4767]: I0317 15:56:05.769809 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcjxq\" (UniqueName: \"kubernetes.io/projected/9d7c1cce-3729-4157-9f4e-275c46c6e436-kube-api-access-rcjxq\") on node \"crc\" DevicePath \"\"" Mar 17 15:56:06 crc kubenswrapper[4767]: I0317 15:56:06.147291 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562716-54ssn" event={"ID":"9d7c1cce-3729-4157-9f4e-275c46c6e436","Type":"ContainerDied","Data":"90456c7cbabec9b1d010036ecdcc2dd909399d193e7b7804743ef4bec3c6c8ce"} Mar 17 15:56:06 crc kubenswrapper[4767]: I0317 15:56:06.147354 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90456c7cbabec9b1d010036ecdcc2dd909399d193e7b7804743ef4bec3c6c8ce" Mar 17 15:56:06 crc kubenswrapper[4767]: I0317 15:56:06.147374 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562716-54ssn" Mar 17 15:56:06 crc kubenswrapper[4767]: I0317 15:56:06.229982 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562710-2bppx"] Mar 17 15:56:06 crc kubenswrapper[4767]: I0317 15:56:06.237375 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562710-2bppx"] Mar 17 15:56:07 crc kubenswrapper[4767]: I0317 15:56:07.487797 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f907dc9-1b30-4ae0-8b07-0566948f13bf" path="/var/lib/kubelet/pods/2f907dc9-1b30-4ae0-8b07-0566948f13bf/volumes" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.805365 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r"] Mar 17 15:56:10 crc kubenswrapper[4767]: E0317 15:56:10.805827 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d7c1cce-3729-4157-9f4e-275c46c6e436" containerName="oc" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.805847 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d7c1cce-3729-4157-9f4e-275c46c6e436" containerName="oc" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.806031 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d7c1cce-3729-4157-9f4e-275c46c6e436" containerName="oc" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.806884 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.814682 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.815266 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.815480 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-mk87w" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.815917 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.815525 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.827345 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r"] Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.866684 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/877f4eda-1ec7-4296-98df-b5ca7a7fa78a-apiservice-cert\") pod \"metallb-operator-controller-manager-5c5f94f4bc-xqf4r\" (UID: \"877f4eda-1ec7-4296-98df-b5ca7a7fa78a\") " pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.867252 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/877f4eda-1ec7-4296-98df-b5ca7a7fa78a-webhook-cert\") pod \"metallb-operator-controller-manager-5c5f94f4bc-xqf4r\" (UID: \"877f4eda-1ec7-4296-98df-b5ca7a7fa78a\") " pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.867341 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4c4h\" (UniqueName: \"kubernetes.io/projected/877f4eda-1ec7-4296-98df-b5ca7a7fa78a-kube-api-access-r4c4h\") pod \"metallb-operator-controller-manager-5c5f94f4bc-xqf4r\" (UID: \"877f4eda-1ec7-4296-98df-b5ca7a7fa78a\") " pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.969588 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/877f4eda-1ec7-4296-98df-b5ca7a7fa78a-apiservice-cert\") pod \"metallb-operator-controller-manager-5c5f94f4bc-xqf4r\" (UID: \"877f4eda-1ec7-4296-98df-b5ca7a7fa78a\") " pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.970273 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/877f4eda-1ec7-4296-98df-b5ca7a7fa78a-webhook-cert\") pod \"metallb-operator-controller-manager-5c5f94f4bc-xqf4r\" (UID: \"877f4eda-1ec7-4296-98df-b5ca7a7fa78a\") " pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.970365 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4c4h\" (UniqueName: \"kubernetes.io/projected/877f4eda-1ec7-4296-98df-b5ca7a7fa78a-kube-api-access-r4c4h\") pod \"metallb-operator-controller-manager-5c5f94f4bc-xqf4r\" (UID: \"877f4eda-1ec7-4296-98df-b5ca7a7fa78a\") " pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.995423 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/877f4eda-1ec7-4296-98df-b5ca7a7fa78a-apiservice-cert\") pod \"metallb-operator-controller-manager-5c5f94f4bc-xqf4r\" (UID: \"877f4eda-1ec7-4296-98df-b5ca7a7fa78a\") " pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:10 crc kubenswrapper[4767]: I0317 15:56:10.995530 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/877f4eda-1ec7-4296-98df-b5ca7a7fa78a-webhook-cert\") pod \"metallb-operator-controller-manager-5c5f94f4bc-xqf4r\" (UID: \"877f4eda-1ec7-4296-98df-b5ca7a7fa78a\") " pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.041100 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4c4h\" (UniqueName: \"kubernetes.io/projected/877f4eda-1ec7-4296-98df-b5ca7a7fa78a-kube-api-access-r4c4h\") pod \"metallb-operator-controller-manager-5c5f94f4bc-xqf4r\" (UID: \"877f4eda-1ec7-4296-98df-b5ca7a7fa78a\") " pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.137832 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d"] Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.139078 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.142102 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.142571 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.154541 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.155660 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d"] Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.173050 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-g6bk8" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.276335 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s75jm\" (UniqueName: \"kubernetes.io/projected/fda1c393-424a-4142-8570-5dc108f6d6f4-kube-api-access-s75jm\") pod \"metallb-operator-webhook-server-5974d9c54d-9rs7d\" (UID: \"fda1c393-424a-4142-8570-5dc108f6d6f4\") " pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.276497 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fda1c393-424a-4142-8570-5dc108f6d6f4-webhook-cert\") pod \"metallb-operator-webhook-server-5974d9c54d-9rs7d\" (UID: \"fda1c393-424a-4142-8570-5dc108f6d6f4\") " pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.276593 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fda1c393-424a-4142-8570-5dc108f6d6f4-apiservice-cert\") pod \"metallb-operator-webhook-server-5974d9c54d-9rs7d\" (UID: \"fda1c393-424a-4142-8570-5dc108f6d6f4\") " pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.386263 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fda1c393-424a-4142-8570-5dc108f6d6f4-apiservice-cert\") pod \"metallb-operator-webhook-server-5974d9c54d-9rs7d\" (UID: \"fda1c393-424a-4142-8570-5dc108f6d6f4\") " pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.386944 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s75jm\" (UniqueName: \"kubernetes.io/projected/fda1c393-424a-4142-8570-5dc108f6d6f4-kube-api-access-s75jm\") pod \"metallb-operator-webhook-server-5974d9c54d-9rs7d\" (UID: \"fda1c393-424a-4142-8570-5dc108f6d6f4\") " pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.387064 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fda1c393-424a-4142-8570-5dc108f6d6f4-webhook-cert\") pod \"metallb-operator-webhook-server-5974d9c54d-9rs7d\" (UID: \"fda1c393-424a-4142-8570-5dc108f6d6f4\") " pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.405634 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fda1c393-424a-4142-8570-5dc108f6d6f4-apiservice-cert\") pod \"metallb-operator-webhook-server-5974d9c54d-9rs7d\" (UID: \"fda1c393-424a-4142-8570-5dc108f6d6f4\") " pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.411081 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fda1c393-424a-4142-8570-5dc108f6d6f4-webhook-cert\") pod \"metallb-operator-webhook-server-5974d9c54d-9rs7d\" (UID: \"fda1c393-424a-4142-8570-5dc108f6d6f4\") " pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.427304 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s75jm\" (UniqueName: \"kubernetes.io/projected/fda1c393-424a-4142-8570-5dc108f6d6f4-kube-api-access-s75jm\") pod \"metallb-operator-webhook-server-5974d9c54d-9rs7d\" (UID: \"fda1c393-424a-4142-8570-5dc108f6d6f4\") " pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.479206 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.764309 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r"] Mar 17 15:56:11 crc kubenswrapper[4767]: I0317 15:56:11.802857 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 15:56:12 crc kubenswrapper[4767]: I0317 15:56:12.281855 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d"] Mar 17 15:56:12 crc kubenswrapper[4767]: W0317 15:56:12.288484 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfda1c393_424a_4142_8570_5dc108f6d6f4.slice/crio-c041949b7333cc0bf9b00cc07927493ff954b8001680208866ca9be246f1494b WatchSource:0}: Error finding container c041949b7333cc0bf9b00cc07927493ff954b8001680208866ca9be246f1494b: Status 404 returned error can't find the container with id c041949b7333cc0bf9b00cc07927493ff954b8001680208866ca9be246f1494b Mar 17 15:56:12 crc kubenswrapper[4767]: I0317 15:56:12.325072 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" event={"ID":"fda1c393-424a-4142-8570-5dc108f6d6f4","Type":"ContainerStarted","Data":"c041949b7333cc0bf9b00cc07927493ff954b8001680208866ca9be246f1494b"} Mar 17 15:56:12 crc kubenswrapper[4767]: I0317 15:56:12.326888 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" event={"ID":"877f4eda-1ec7-4296-98df-b5ca7a7fa78a","Type":"ContainerStarted","Data":"61a2b733a7cb2eb441955e530bc96d2b63e580fc1b122e676e024a9aa3f398fe"} Mar 17 15:56:22 crc kubenswrapper[4767]: I0317 15:56:22.471273 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" event={"ID":"fda1c393-424a-4142-8570-5dc108f6d6f4","Type":"ContainerStarted","Data":"f8be811582075c48d993dad19d81c0727dfffbb03d827ac409ec6fe280626d48"} Mar 17 15:56:22 crc kubenswrapper[4767]: I0317 15:56:22.472187 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:22 crc kubenswrapper[4767]: I0317 15:56:22.473714 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" event={"ID":"877f4eda-1ec7-4296-98df-b5ca7a7fa78a","Type":"ContainerStarted","Data":"b8a216894dbce2190c35e4629181d1975a45c2bddb4479f3c33d94755e92e1a2"} Mar 17 15:56:22 crc kubenswrapper[4767]: I0317 15:56:22.473899 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:22 crc kubenswrapper[4767]: I0317 15:56:22.524096 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" podStartSLOduration=2.296609029 podStartE2EDuration="11.524060058s" podCreationTimestamp="2026-03-17 15:56:11 +0000 UTC" firstStartedPulling="2026-03-17 15:56:12.293324153 +0000 UTC m=+1163.706640200" lastFinishedPulling="2026-03-17 15:56:21.520775182 +0000 UTC m=+1172.934091229" observedRunningTime="2026-03-17 15:56:22.514285215 +0000 UTC m=+1173.927601292" watchObservedRunningTime="2026-03-17 15:56:22.524060058 +0000 UTC m=+1173.937376105" Mar 17 15:56:22 crc kubenswrapper[4767]: I0317 15:56:22.551475 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" podStartSLOduration=2.859191404 podStartE2EDuration="12.551447711s" podCreationTimestamp="2026-03-17 15:56:10 +0000 UTC" firstStartedPulling="2026-03-17 15:56:11.802607005 +0000 UTC m=+1163.215923052" lastFinishedPulling="2026-03-17 15:56:21.494863312 +0000 UTC m=+1172.908179359" observedRunningTime="2026-03-17 15:56:22.545712255 +0000 UTC m=+1173.959028312" watchObservedRunningTime="2026-03-17 15:56:22.551447711 +0000 UTC m=+1173.964763758" Mar 17 15:56:31 crc kubenswrapper[4767]: I0317 15:56:31.665053 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 15:56:33 crc kubenswrapper[4767]: I0317 15:56:33.038163 4767 scope.go:117] "RemoveContainer" containerID="32bbcb6ddd2c3090708f57b91d7c71398f21718cabb4c5c0710bdedff7bc91a2" Mar 17 15:56:34 crc kubenswrapper[4767]: I0317 15:56:34.167021 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:56:34 crc kubenswrapper[4767]: I0317 15:56:34.167652 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:56:51 crc kubenswrapper[4767]: I0317 15:56:51.157688 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.070960 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-wj8xp"] Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.076695 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.081740 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.082016 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.093002 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-9zbbf" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.093583 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7"] Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.094900 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.102264 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.119050 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7"] Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.177116 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-wjz85"] Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.182896 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.187721 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.188030 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.188147 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.194393 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-qhk7n" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.214133 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-nchz8"] Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.216013 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.220329 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.250839 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e5da4ccd-4ecb-4929-8974-0a31d018204d-frr-startup\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.250946 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-metrics\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.250972 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nkxl\" (UniqueName: \"kubernetes.io/projected/07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec-kube-api-access-8nkxl\") pod \"frr-k8s-webhook-server-bcc4b6f68-khzr7\" (UID: \"07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.251023 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nqtk\" (UniqueName: \"kubernetes.io/projected/e5da4ccd-4ecb-4929-8974-0a31d018204d-kube-api-access-8nqtk\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.251050 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-khzr7\" (UID: \"07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.251072 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5da4ccd-4ecb-4929-8974-0a31d018204d-metrics-certs\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.251105 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-reloader\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.251120 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-frr-conf\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.251156 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-frr-sockets\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.256157 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-nchz8"] Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.353275 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-metrics-certs\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.353358 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9wl8\" (UniqueName: \"kubernetes.io/projected/1cfcd066-3d8a-431e-a895-a257a3e1baef-kube-api-access-b9wl8\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.353400 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nqtk\" (UniqueName: \"kubernetes.io/projected/e5da4ccd-4ecb-4929-8974-0a31d018204d-kube-api-access-8nqtk\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.353428 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-khzr7\" (UID: \"07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.353579 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5da4ccd-4ecb-4929-8974-0a31d018204d-metrics-certs\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.353804 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2jpz\" (UniqueName: \"kubernetes.io/projected/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-kube-api-access-r2jpz\") pod \"controller-7bb4cc7c98-nchz8\" (UID: \"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4\") " pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.353920 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-reloader\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.353954 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-frr-conf\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.354506 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-frr-conf\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.354555 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-reloader\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.354649 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-frr-sockets\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.354938 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-frr-sockets\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.354997 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.355054 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e5da4ccd-4ecb-4929-8974-0a31d018204d-frr-startup\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.355081 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-cert\") pod \"controller-7bb4cc7c98-nchz8\" (UID: \"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4\") " pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.355223 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1cfcd066-3d8a-431e-a895-a257a3e1baef-metallb-excludel2\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.355264 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-metrics\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.355300 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nkxl\" (UniqueName: \"kubernetes.io/projected/07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec-kube-api-access-8nkxl\") pod \"frr-k8s-webhook-server-bcc4b6f68-khzr7\" (UID: \"07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.355360 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-metrics-certs\") pod \"controller-7bb4cc7c98-nchz8\" (UID: \"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4\") " pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.355902 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e5da4ccd-4ecb-4929-8974-0a31d018204d-metrics\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.360878 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5da4ccd-4ecb-4929-8974-0a31d018204d-metrics-certs\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.361873 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e5da4ccd-4ecb-4929-8974-0a31d018204d-frr-startup\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.362400 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-khzr7\" (UID: \"07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.375882 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nqtk\" (UniqueName: \"kubernetes.io/projected/e5da4ccd-4ecb-4929-8974-0a31d018204d-kube-api-access-8nqtk\") pod \"frr-k8s-wj8xp\" (UID: \"e5da4ccd-4ecb-4929-8974-0a31d018204d\") " pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.386980 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nkxl\" (UniqueName: \"kubernetes.io/projected/07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec-kube-api-access-8nkxl\") pod \"frr-k8s-webhook-server-bcc4b6f68-khzr7\" (UID: \"07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.419009 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.434952 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.456805 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2jpz\" (UniqueName: \"kubernetes.io/projected/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-kube-api-access-r2jpz\") pod \"controller-7bb4cc7c98-nchz8\" (UID: \"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4\") " pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.457110 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.457283 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-cert\") pod \"controller-7bb4cc7c98-nchz8\" (UID: \"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4\") " pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.457434 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1cfcd066-3d8a-431e-a895-a257a3e1baef-metallb-excludel2\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.457556 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-metrics-certs\") pod \"controller-7bb4cc7c98-nchz8\" (UID: \"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4\") " pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.457649 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-metrics-certs\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.457758 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9wl8\" (UniqueName: \"kubernetes.io/projected/1cfcd066-3d8a-431e-a895-a257a3e1baef-kube-api-access-b9wl8\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: E0317 15:56:52.458158 4767 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Mar 17 15:56:52 crc kubenswrapper[4767]: E0317 15:56:52.458364 4767 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 17 15:56:52 crc kubenswrapper[4767]: E0317 15:56:52.458379 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-metrics-certs podName:b978bf52-fdb5-4863-a5bf-e2aec8ab08a4 nodeName:}" failed. No retries permitted until 2026-03-17 15:56:52.95833513 +0000 UTC m=+1204.371651287 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-metrics-certs") pod "controller-7bb4cc7c98-nchz8" (UID: "b978bf52-fdb5-4863-a5bf-e2aec8ab08a4") : secret "controller-certs-secret" not found Mar 17 15:56:52 crc kubenswrapper[4767]: E0317 15:56:52.458484 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist podName:1cfcd066-3d8a-431e-a895-a257a3e1baef nodeName:}" failed. No retries permitted until 2026-03-17 15:56:52.958453633 +0000 UTC m=+1204.371769750 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist") pod "speaker-wjz85" (UID: "1cfcd066-3d8a-431e-a895-a257a3e1baef") : secret "metallb-memberlist" not found Mar 17 15:56:52 crc kubenswrapper[4767]: E0317 15:56:52.458854 4767 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.458885 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1cfcd066-3d8a-431e-a895-a257a3e1baef-metallb-excludel2\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: E0317 15:56:52.458892 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-metrics-certs podName:1cfcd066-3d8a-431e-a895-a257a3e1baef nodeName:}" failed. No retries permitted until 2026-03-17 15:56:52.958883515 +0000 UTC m=+1204.372199562 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-metrics-certs") pod "speaker-wjz85" (UID: "1cfcd066-3d8a-431e-a895-a257a3e1baef") : secret "speaker-certs-secret" not found Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.463395 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.483670 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-cert\") pod \"controller-7bb4cc7c98-nchz8\" (UID: \"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4\") " pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.484383 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9wl8\" (UniqueName: \"kubernetes.io/projected/1cfcd066-3d8a-431e-a895-a257a3e1baef-kube-api-access-b9wl8\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.485082 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2jpz\" (UniqueName: \"kubernetes.io/projected/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-kube-api-access-r2jpz\") pod \"controller-7bb4cc7c98-nchz8\" (UID: \"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4\") " pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.927229 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7"] Mar 17 15:56:52 crc kubenswrapper[4767]: W0317 15:56:52.934431 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07d5db0e_27a0_43ee_a8a7_54c27bf6f9ec.slice/crio-2ea8a6213bfaa0686ad0ddce298013ee690292fe40148e53e40aa58e7615c56b WatchSource:0}: Error finding container 2ea8a6213bfaa0686ad0ddce298013ee690292fe40148e53e40aa58e7615c56b: Status 404 returned error can't find the container with id 2ea8a6213bfaa0686ad0ddce298013ee690292fe40148e53e40aa58e7615c56b Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.969011 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.969153 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-metrics-certs\") pod \"controller-7bb4cc7c98-nchz8\" (UID: \"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4\") " pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.969206 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-metrics-certs\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: E0317 15:56:52.969837 4767 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 17 15:56:52 crc kubenswrapper[4767]: E0317 15:56:52.969972 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist podName:1cfcd066-3d8a-431e-a895-a257a3e1baef nodeName:}" failed. No retries permitted until 2026-03-17 15:56:53.969943771 +0000 UTC m=+1205.383259818 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist") pod "speaker-wjz85" (UID: "1cfcd066-3d8a-431e-a895-a257a3e1baef") : secret "metallb-memberlist" not found Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.976044 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-metrics-certs\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:52 crc kubenswrapper[4767]: I0317 15:56:52.976224 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b978bf52-fdb5-4863-a5bf-e2aec8ab08a4-metrics-certs\") pod \"controller-7bb4cc7c98-nchz8\" (UID: \"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4\") " pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:53 crc kubenswrapper[4767]: I0317 15:56:53.126993 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" event={"ID":"07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec","Type":"ContainerStarted","Data":"2ea8a6213bfaa0686ad0ddce298013ee690292fe40148e53e40aa58e7615c56b"} Mar 17 15:56:53 crc kubenswrapper[4767]: I0317 15:56:53.133164 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerStarted","Data":"cce3d228169de2c63817fbdf48c13babc425e5a85762b351247a4fb4b1023154"} Mar 17 15:56:53 crc kubenswrapper[4767]: I0317 15:56:53.151114 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:53 crc kubenswrapper[4767]: I0317 15:56:53.749677 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-nchz8"] Mar 17 15:56:54 crc kubenswrapper[4767]: I0317 15:56:54.008916 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:54 crc kubenswrapper[4767]: E0317 15:56:54.009126 4767 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 17 15:56:54 crc kubenswrapper[4767]: E0317 15:56:54.009257 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist podName:1cfcd066-3d8a-431e-a895-a257a3e1baef nodeName:}" failed. No retries permitted until 2026-03-17 15:56:56.009224738 +0000 UTC m=+1207.422540785 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist") pod "speaker-wjz85" (UID: "1cfcd066-3d8a-431e-a895-a257a3e1baef") : secret "metallb-memberlist" not found Mar 17 15:56:54 crc kubenswrapper[4767]: I0317 15:56:54.155566 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-nchz8" event={"ID":"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4","Type":"ContainerStarted","Data":"69e01d6fa79286a2a561b8be795c0b0030b4cd99ae54c7282a923a227b836b42"} Mar 17 15:56:54 crc kubenswrapper[4767]: I0317 15:56:54.155694 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-nchz8" event={"ID":"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4","Type":"ContainerStarted","Data":"b9796b6f22895c77e87432d47b6bec018783c32c4048124d1c8971f84565c5d6"} Mar 17 15:56:55 crc kubenswrapper[4767]: I0317 15:56:55.182359 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-nchz8" event={"ID":"b978bf52-fdb5-4863-a5bf-e2aec8ab08a4","Type":"ContainerStarted","Data":"7ac70f49dfca9e0eb75703f9f938ae7a39853b1f53a8243ea2b9be8b951af9db"} Mar 17 15:56:55 crc kubenswrapper[4767]: I0317 15:56:55.184432 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:56:55 crc kubenswrapper[4767]: I0317 15:56:55.224213 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-nchz8" podStartSLOduration=3.224158571 podStartE2EDuration="3.224158571s" podCreationTimestamp="2026-03-17 15:56:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:56:55.219039973 +0000 UTC m=+1206.632356030" watchObservedRunningTime="2026-03-17 15:56:55.224158571 +0000 UTC m=+1206.637474618" Mar 17 15:56:56 crc kubenswrapper[4767]: I0317 15:56:56.067003 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:56 crc kubenswrapper[4767]: I0317 15:56:56.079121 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1cfcd066-3d8a-431e-a895-a257a3e1baef-memberlist\") pod \"speaker-wjz85\" (UID: \"1cfcd066-3d8a-431e-a895-a257a3e1baef\") " pod="metallb-system/speaker-wjz85" Mar 17 15:56:56 crc kubenswrapper[4767]: I0317 15:56:56.115831 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wjz85" Mar 17 15:56:56 crc kubenswrapper[4767]: I0317 15:56:56.226007 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wjz85" event={"ID":"1cfcd066-3d8a-431e-a895-a257a3e1baef","Type":"ContainerStarted","Data":"1da5febdf78b3ffe0c99741653edca272eb999eb555743a9141669a10b014053"} Mar 17 15:56:57 crc kubenswrapper[4767]: I0317 15:56:57.238895 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wjz85" event={"ID":"1cfcd066-3d8a-431e-a895-a257a3e1baef","Type":"ContainerStarted","Data":"ba261f78d850e6e435383d06492c10b028f053d4277ea1d45eeb1556d3f34f41"} Mar 17 15:56:57 crc kubenswrapper[4767]: I0317 15:56:57.239603 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wjz85" event={"ID":"1cfcd066-3d8a-431e-a895-a257a3e1baef","Type":"ContainerStarted","Data":"e809b31439dc0f3d987c0a5d250a1921743dd5246b80134e8aa65b80a195605d"} Mar 17 15:56:57 crc kubenswrapper[4767]: I0317 15:56:57.421011 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-wjz85" podStartSLOduration=5.42098646 podStartE2EDuration="5.42098646s" podCreationTimestamp="2026-03-17 15:56:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:56:57.2842184 +0000 UTC m=+1208.697534447" watchObservedRunningTime="2026-03-17 15:56:57.42098646 +0000 UTC m=+1208.834302507" Mar 17 15:56:58 crc kubenswrapper[4767]: I0317 15:56:58.247902 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-wjz85" Mar 17 15:57:03 crc kubenswrapper[4767]: I0317 15:57:03.157573 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-nchz8" Mar 17 15:57:04 crc kubenswrapper[4767]: I0317 15:57:04.166930 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:57:04 crc kubenswrapper[4767]: I0317 15:57:04.167601 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:57:05 crc kubenswrapper[4767]: I0317 15:57:05.938354 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" event={"ID":"07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec","Type":"ContainerStarted","Data":"419cc940f05664f6f7e89224b92ed5c9bd7ed965f7d6d4de25d564d24a720843"} Mar 17 15:57:05 crc kubenswrapper[4767]: I0317 15:57:05.938658 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 15:57:05 crc kubenswrapper[4767]: I0317 15:57:05.945245 4767 generic.go:334] "Generic (PLEG): container finished" podID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerID="f9c0d4ea207c7c2506da0481c85bdd927721bfd1d5a5f1c59322c043d0a11eff" exitCode=0 Mar 17 15:57:05 crc kubenswrapper[4767]: I0317 15:57:05.945297 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerDied","Data":"f9c0d4ea207c7c2506da0481c85bdd927721bfd1d5a5f1c59322c043d0a11eff"} Mar 17 15:57:05 crc kubenswrapper[4767]: I0317 15:57:05.965963 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" podStartSLOduration=2.673077447 podStartE2EDuration="14.965931468s" podCreationTimestamp="2026-03-17 15:56:51 +0000 UTC" firstStartedPulling="2026-03-17 15:56:52.937727798 +0000 UTC m=+1204.351043845" lastFinishedPulling="2026-03-17 15:57:05.230581819 +0000 UTC m=+1216.643897866" observedRunningTime="2026-03-17 15:57:05.957312239 +0000 UTC m=+1217.370628306" watchObservedRunningTime="2026-03-17 15:57:05.965931468 +0000 UTC m=+1217.379247525" Mar 17 15:57:06 crc kubenswrapper[4767]: I0317 15:57:06.124401 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-wjz85" Mar 17 15:57:06 crc kubenswrapper[4767]: I0317 15:57:06.957425 4767 generic.go:334] "Generic (PLEG): container finished" podID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerID="7e57be0c7af829bc09afed0a9d84fc198e0c0f8545e24351bea9c86e752bec21" exitCode=0 Mar 17 15:57:06 crc kubenswrapper[4767]: I0317 15:57:06.957543 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerDied","Data":"7e57be0c7af829bc09afed0a9d84fc198e0c0f8545e24351bea9c86e752bec21"} Mar 17 15:57:07 crc kubenswrapper[4767]: I0317 15:57:07.969942 4767 generic.go:334] "Generic (PLEG): container finished" podID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerID="eca815fda085055e83f5d3c23b868add1a9fbe7662d65a7752ab47afc46a79f7" exitCode=0 Mar 17 15:57:07 crc kubenswrapper[4767]: I0317 15:57:07.970024 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerDied","Data":"eca815fda085055e83f5d3c23b868add1a9fbe7662d65a7752ab47afc46a79f7"} Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:08.992940 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerStarted","Data":"422f31bf6258404fb472585dacded8860e835d06d8be14a45a3661c844377a0b"} Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:08.993873 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerStarted","Data":"0d62329195f057cb89605f819910baaad4fb56deadc86d63f22640f71d9f3d27"} Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:08.993884 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerStarted","Data":"73b35f1d53bbc898405c6748eb7e07c6cd4c0b66508ad06888ba4c096c17b6f4"} Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:08.993893 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerStarted","Data":"fbfb64eda9a1857b0e600f098e323af2c36d82414c8ce1e3bffb3520a89ba819"} Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:08.993920 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerStarted","Data":"b3c9ce2a1c6b819b46adecf1b0af9f33b68e25920278ace693bc6bd14479bdf5"} Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:09.109788 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-j4zdw"] Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:09.111773 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-j4zdw" Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:09.117673 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:09.117674 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-tphpc" Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:09.129859 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:09.143845 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-j4zdw"] Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:09.194669 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq8tj\" (UniqueName: \"kubernetes.io/projected/03cf6362-f451-427f-b89f-3eee61667d77-kube-api-access-pq8tj\") pod \"openstack-operator-index-j4zdw\" (UID: \"03cf6362-f451-427f-b89f-3eee61667d77\") " pod="openstack-operators/openstack-operator-index-j4zdw" Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:09.297469 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq8tj\" (UniqueName: \"kubernetes.io/projected/03cf6362-f451-427f-b89f-3eee61667d77-kube-api-access-pq8tj\") pod \"openstack-operator-index-j4zdw\" (UID: \"03cf6362-f451-427f-b89f-3eee61667d77\") " pod="openstack-operators/openstack-operator-index-j4zdw" Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:09.321200 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq8tj\" (UniqueName: \"kubernetes.io/projected/03cf6362-f451-427f-b89f-3eee61667d77-kube-api-access-pq8tj\") pod \"openstack-operator-index-j4zdw\" (UID: \"03cf6362-f451-427f-b89f-3eee61667d77\") " pod="openstack-operators/openstack-operator-index-j4zdw" Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:09.460979 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-j4zdw" Mar 17 15:57:09 crc kubenswrapper[4767]: I0317 15:57:09.926819 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-j4zdw"] Mar 17 15:57:10 crc kubenswrapper[4767]: I0317 15:57:10.005754 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-j4zdw" event={"ID":"03cf6362-f451-427f-b89f-3eee61667d77","Type":"ContainerStarted","Data":"7ad086fdd4a8c079573cd1f49de5e6800cf86cad005aa46737263ef063ad1077"} Mar 17 15:57:10 crc kubenswrapper[4767]: I0317 15:57:10.011796 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerStarted","Data":"f05b2b7bfa9cb666cdcea924271f1c3896e5f42e159ebcec89153185128785a4"} Mar 17 15:57:10 crc kubenswrapper[4767]: I0317 15:57:10.012946 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:57:12 crc kubenswrapper[4767]: I0317 15:57:12.260153 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-wj8xp" podStartSLOduration=8.681227654 podStartE2EDuration="21.260118916s" podCreationTimestamp="2026-03-17 15:56:51 +0000 UTC" firstStartedPulling="2026-03-17 15:56:52.671182232 +0000 UTC m=+1204.084498279" lastFinishedPulling="2026-03-17 15:57:05.250073494 +0000 UTC m=+1216.663389541" observedRunningTime="2026-03-17 15:57:10.056023526 +0000 UTC m=+1221.469339593" watchObservedRunningTime="2026-03-17 15:57:12.260118916 +0000 UTC m=+1223.673434983" Mar 17 15:57:12 crc kubenswrapper[4767]: I0317 15:57:12.263454 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-j4zdw"] Mar 17 15:57:12 crc kubenswrapper[4767]: I0317 15:57:12.419716 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:57:12 crc kubenswrapper[4767]: I0317 15:57:12.457533 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:57:12 crc kubenswrapper[4767]: I0317 15:57:12.878409 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-mdghs"] Mar 17 15:57:12 crc kubenswrapper[4767]: I0317 15:57:12.880191 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mdghs" Mar 17 15:57:12 crc kubenswrapper[4767]: I0317 15:57:12.888940 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mdghs"] Mar 17 15:57:13 crc kubenswrapper[4767]: I0317 15:57:13.039947 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b5s8\" (UniqueName: \"kubernetes.io/projected/571d124c-f5cb-4c5e-bfd7-3f56836ce800-kube-api-access-4b5s8\") pod \"openstack-operator-index-mdghs\" (UID: \"571d124c-f5cb-4c5e-bfd7-3f56836ce800\") " pod="openstack-operators/openstack-operator-index-mdghs" Mar 17 15:57:13 crc kubenswrapper[4767]: I0317 15:57:13.141568 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b5s8\" (UniqueName: \"kubernetes.io/projected/571d124c-f5cb-4c5e-bfd7-3f56836ce800-kube-api-access-4b5s8\") pod \"openstack-operator-index-mdghs\" (UID: \"571d124c-f5cb-4c5e-bfd7-3f56836ce800\") " pod="openstack-operators/openstack-operator-index-mdghs" Mar 17 15:57:13 crc kubenswrapper[4767]: I0317 15:57:13.163879 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b5s8\" (UniqueName: \"kubernetes.io/projected/571d124c-f5cb-4c5e-bfd7-3f56836ce800-kube-api-access-4b5s8\") pod \"openstack-operator-index-mdghs\" (UID: \"571d124c-f5cb-4c5e-bfd7-3f56836ce800\") " pod="openstack-operators/openstack-operator-index-mdghs" Mar 17 15:57:13 crc kubenswrapper[4767]: I0317 15:57:13.220680 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mdghs" Mar 17 15:57:13 crc kubenswrapper[4767]: I0317 15:57:13.941720 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mdghs"] Mar 17 15:57:14 crc kubenswrapper[4767]: I0317 15:57:14.053564 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mdghs" event={"ID":"571d124c-f5cb-4c5e-bfd7-3f56836ce800","Type":"ContainerStarted","Data":"52890b2061a8f608c745cf9ec34631a475e9974bfbe1e122681a307173fb8348"} Mar 17 15:57:14 crc kubenswrapper[4767]: I0317 15:57:14.055966 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-j4zdw" event={"ID":"03cf6362-f451-427f-b89f-3eee61667d77","Type":"ContainerStarted","Data":"bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d"} Mar 17 15:57:14 crc kubenswrapper[4767]: I0317 15:57:14.056235 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-j4zdw" podUID="03cf6362-f451-427f-b89f-3eee61667d77" containerName="registry-server" containerID="cri-o://bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d" gracePeriod=2 Mar 17 15:57:14 crc kubenswrapper[4767]: I0317 15:57:14.089650 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-j4zdw" podStartSLOduration=2.039360832 podStartE2EDuration="5.089610279s" podCreationTimestamp="2026-03-17 15:57:09 +0000 UTC" firstStartedPulling="2026-03-17 15:57:09.958239805 +0000 UTC m=+1221.371555852" lastFinishedPulling="2026-03-17 15:57:13.008489252 +0000 UTC m=+1224.421805299" observedRunningTime="2026-03-17 15:57:14.08133425 +0000 UTC m=+1225.494650387" watchObservedRunningTime="2026-03-17 15:57:14.089610279 +0000 UTC m=+1225.502926326" Mar 17 15:57:14 crc kubenswrapper[4767]: I0317 15:57:14.925769 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-j4zdw" Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.067577 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mdghs" event={"ID":"571d124c-f5cb-4c5e-bfd7-3f56836ce800","Type":"ContainerStarted","Data":"1fbd283d482e1a64b19ee94a0782de04e8fd3a5d3b5e4ecc080837026bc5fbc8"} Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.069288 4767 generic.go:334] "Generic (PLEG): container finished" podID="03cf6362-f451-427f-b89f-3eee61667d77" containerID="bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d" exitCode=0 Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.069329 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-j4zdw" event={"ID":"03cf6362-f451-427f-b89f-3eee61667d77","Type":"ContainerDied","Data":"bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d"} Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.069360 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-j4zdw" event={"ID":"03cf6362-f451-427f-b89f-3eee61667d77","Type":"ContainerDied","Data":"7ad086fdd4a8c079573cd1f49de5e6800cf86cad005aa46737263ef063ad1077"} Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.069389 4767 scope.go:117] "RemoveContainer" containerID="bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d" Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.069586 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-j4zdw" Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.097477 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-mdghs" podStartSLOduration=3.037694167 podStartE2EDuration="3.097437997s" podCreationTimestamp="2026-03-17 15:57:12 +0000 UTC" firstStartedPulling="2026-03-17 15:57:13.960581354 +0000 UTC m=+1225.373897401" lastFinishedPulling="2026-03-17 15:57:14.020325164 +0000 UTC m=+1225.433641231" observedRunningTime="2026-03-17 15:57:15.086622684 +0000 UTC m=+1226.499938741" watchObservedRunningTime="2026-03-17 15:57:15.097437997 +0000 UTC m=+1226.510754044" Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.104266 4767 scope.go:117] "RemoveContainer" containerID="bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d" Mar 17 15:57:15 crc kubenswrapper[4767]: E0317 15:57:15.105103 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d\": container with ID starting with bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d not found: ID does not exist" containerID="bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d" Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.105144 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d"} err="failed to get container status \"bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d\": rpc error: code = NotFound desc = could not find container \"bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d\": container with ID starting with bb1842cc381a3887f2d3de6e67ac439431643b878400de113388ea72d243dd7d not found: ID does not exist" Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.124481 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq8tj\" (UniqueName: \"kubernetes.io/projected/03cf6362-f451-427f-b89f-3eee61667d77-kube-api-access-pq8tj\") pod \"03cf6362-f451-427f-b89f-3eee61667d77\" (UID: \"03cf6362-f451-427f-b89f-3eee61667d77\") " Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.134430 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03cf6362-f451-427f-b89f-3eee61667d77-kube-api-access-pq8tj" (OuterVolumeSpecName: "kube-api-access-pq8tj") pod "03cf6362-f451-427f-b89f-3eee61667d77" (UID: "03cf6362-f451-427f-b89f-3eee61667d77"). InnerVolumeSpecName "kube-api-access-pq8tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.267354 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq8tj\" (UniqueName: \"kubernetes.io/projected/03cf6362-f451-427f-b89f-3eee61667d77-kube-api-access-pq8tj\") on node \"crc\" DevicePath \"\"" Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.412280 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-j4zdw"] Mar 17 15:57:15 crc kubenswrapper[4767]: I0317 15:57:15.423629 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-j4zdw"] Mar 17 15:57:17 crc kubenswrapper[4767]: I0317 15:57:17.366054 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03cf6362-f451-427f-b89f-3eee61667d77" path="/var/lib/kubelet/pods/03cf6362-f451-427f-b89f-3eee61667d77/volumes" Mar 17 15:57:22 crc kubenswrapper[4767]: I0317 15:57:22.427296 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-wj8xp" Mar 17 15:57:22 crc kubenswrapper[4767]: I0317 15:57:22.443201 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 15:57:23 crc kubenswrapper[4767]: I0317 15:57:23.221612 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-mdghs" Mar 17 15:57:23 crc kubenswrapper[4767]: I0317 15:57:23.222328 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-mdghs" Mar 17 15:57:23 crc kubenswrapper[4767]: I0317 15:57:23.261228 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-mdghs" Mar 17 15:57:24 crc kubenswrapper[4767]: I0317 15:57:24.194086 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-mdghs" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.108586 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f"] Mar 17 15:57:30 crc kubenswrapper[4767]: E0317 15:57:30.109981 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03cf6362-f451-427f-b89f-3eee61667d77" containerName="registry-server" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.109999 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="03cf6362-f451-427f-b89f-3eee61667d77" containerName="registry-server" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.110184 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="03cf6362-f451-427f-b89f-3eee61667d77" containerName="registry-server" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.111573 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.116667 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7vxtd" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.121292 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f"] Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.257836 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp5hm\" (UniqueName: \"kubernetes.io/projected/36731840-c502-4b79-a9d9-3dcede160c19-kube-api-access-rp5hm\") pod \"059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.257904 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-bundle\") pod \"059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.257927 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-util\") pod \"059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.360115 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp5hm\" (UniqueName: \"kubernetes.io/projected/36731840-c502-4b79-a9d9-3dcede160c19-kube-api-access-rp5hm\") pod \"059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.360206 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-bundle\") pod \"059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.360240 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-util\") pod \"059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.360807 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-bundle\") pod \"059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.360975 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-util\") pod \"059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.383008 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp5hm\" (UniqueName: \"kubernetes.io/projected/36731840-c502-4b79-a9d9-3dcede160c19-kube-api-access-rp5hm\") pod \"059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.467681 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:30 crc kubenswrapper[4767]: I0317 15:57:30.952570 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f"] Mar 17 15:57:30 crc kubenswrapper[4767]: W0317 15:57:30.961039 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36731840_c502_4b79_a9d9_3dcede160c19.slice/crio-74fceb8531be0d78aac9f48f666484d865a3566e9f63440b2f113b21aeebce60 WatchSource:0}: Error finding container 74fceb8531be0d78aac9f48f666484d865a3566e9f63440b2f113b21aeebce60: Status 404 returned error can't find the container with id 74fceb8531be0d78aac9f48f666484d865a3566e9f63440b2f113b21aeebce60 Mar 17 15:57:31 crc kubenswrapper[4767]: I0317 15:57:31.230737 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" event={"ID":"36731840-c502-4b79-a9d9-3dcede160c19","Type":"ContainerStarted","Data":"b72895a59f8d5b32c8041a8bf55b16d1b55a5070dd8ee5ac05344c3c7905942e"} Mar 17 15:57:31 crc kubenswrapper[4767]: I0317 15:57:31.230797 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" event={"ID":"36731840-c502-4b79-a9d9-3dcede160c19","Type":"ContainerStarted","Data":"74fceb8531be0d78aac9f48f666484d865a3566e9f63440b2f113b21aeebce60"} Mar 17 15:57:32 crc kubenswrapper[4767]: I0317 15:57:32.245471 4767 generic.go:334] "Generic (PLEG): container finished" podID="36731840-c502-4b79-a9d9-3dcede160c19" containerID="b72895a59f8d5b32c8041a8bf55b16d1b55a5070dd8ee5ac05344c3c7905942e" exitCode=0 Mar 17 15:57:32 crc kubenswrapper[4767]: I0317 15:57:32.245599 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" event={"ID":"36731840-c502-4b79-a9d9-3dcede160c19","Type":"ContainerDied","Data":"b72895a59f8d5b32c8041a8bf55b16d1b55a5070dd8ee5ac05344c3c7905942e"} Mar 17 15:57:33 crc kubenswrapper[4767]: I0317 15:57:33.260782 4767 generic.go:334] "Generic (PLEG): container finished" podID="36731840-c502-4b79-a9d9-3dcede160c19" containerID="55433529d740e2e030b95ae394c596b58e279a79c0f9911f524e48ade2eebecc" exitCode=0 Mar 17 15:57:33 crc kubenswrapper[4767]: I0317 15:57:33.261657 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" event={"ID":"36731840-c502-4b79-a9d9-3dcede160c19","Type":"ContainerDied","Data":"55433529d740e2e030b95ae394c596b58e279a79c0f9911f524e48ade2eebecc"} Mar 17 15:57:34 crc kubenswrapper[4767]: I0317 15:57:34.167665 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:57:34 crc kubenswrapper[4767]: I0317 15:57:34.167798 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:57:34 crc kubenswrapper[4767]: I0317 15:57:34.168128 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 15:57:34 crc kubenswrapper[4767]: I0317 15:57:34.169304 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c0a59e1cd71c114bacc3d0ebc2a7a84e16351d4b20330e3bb7479cc322816c8b"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 15:57:34 crc kubenswrapper[4767]: I0317 15:57:34.169377 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://c0a59e1cd71c114bacc3d0ebc2a7a84e16351d4b20330e3bb7479cc322816c8b" gracePeriod=600 Mar 17 15:57:34 crc kubenswrapper[4767]: I0317 15:57:34.288231 4767 generic.go:334] "Generic (PLEG): container finished" podID="36731840-c502-4b79-a9d9-3dcede160c19" containerID="e3237abcc38900dceabdff7fa2743024d046a1f508a18d389b4affafb009f0f7" exitCode=0 Mar 17 15:57:34 crc kubenswrapper[4767]: I0317 15:57:34.288752 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" event={"ID":"36731840-c502-4b79-a9d9-3dcede160c19","Type":"ContainerDied","Data":"e3237abcc38900dceabdff7fa2743024d046a1f508a18d389b4affafb009f0f7"} Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.302659 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="c0a59e1cd71c114bacc3d0ebc2a7a84e16351d4b20330e3bb7479cc322816c8b" exitCode=0 Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.302716 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"c0a59e1cd71c114bacc3d0ebc2a7a84e16351d4b20330e3bb7479cc322816c8b"} Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.303327 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"53b1541419163e416532d56412b5286cd823f9a0b2d0b4b97086b6642dba72af"} Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.303378 4767 scope.go:117] "RemoveContainer" containerID="265b2d9bdc4eb65af73c1e843186a34dfa697843aaf572fda7ed43575c18359a" Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.678979 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.770948 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-bundle\") pod \"36731840-c502-4b79-a9d9-3dcede160c19\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.771297 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-util\") pod \"36731840-c502-4b79-a9d9-3dcede160c19\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.771416 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp5hm\" (UniqueName: \"kubernetes.io/projected/36731840-c502-4b79-a9d9-3dcede160c19-kube-api-access-rp5hm\") pod \"36731840-c502-4b79-a9d9-3dcede160c19\" (UID: \"36731840-c502-4b79-a9d9-3dcede160c19\") " Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.772265 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-bundle" (OuterVolumeSpecName: "bundle") pod "36731840-c502-4b79-a9d9-3dcede160c19" (UID: "36731840-c502-4b79-a9d9-3dcede160c19"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.785644 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36731840-c502-4b79-a9d9-3dcede160c19-kube-api-access-rp5hm" (OuterVolumeSpecName: "kube-api-access-rp5hm") pod "36731840-c502-4b79-a9d9-3dcede160c19" (UID: "36731840-c502-4b79-a9d9-3dcede160c19"). InnerVolumeSpecName "kube-api-access-rp5hm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.787593 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-util" (OuterVolumeSpecName: "util") pod "36731840-c502-4b79-a9d9-3dcede160c19" (UID: "36731840-c502-4b79-a9d9-3dcede160c19"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.873785 4767 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-util\") on node \"crc\" DevicePath \"\"" Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.873846 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp5hm\" (UniqueName: \"kubernetes.io/projected/36731840-c502-4b79-a9d9-3dcede160c19-kube-api-access-rp5hm\") on node \"crc\" DevicePath \"\"" Mar 17 15:57:35 crc kubenswrapper[4767]: I0317 15:57:35.873863 4767 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36731840-c502-4b79-a9d9-3dcede160c19-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 15:57:36 crc kubenswrapper[4767]: I0317 15:57:36.316623 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" Mar 17 15:57:36 crc kubenswrapper[4767]: I0317 15:57:36.316617 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/059e3a718a15b950b608703be204187afb33cb26a54fc3cd5c44fecff1nkg6f" event={"ID":"36731840-c502-4b79-a9d9-3dcede160c19","Type":"ContainerDied","Data":"74fceb8531be0d78aac9f48f666484d865a3566e9f63440b2f113b21aeebce60"} Mar 17 15:57:36 crc kubenswrapper[4767]: I0317 15:57:36.317632 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74fceb8531be0d78aac9f48f666484d865a3566e9f63440b2f113b21aeebce60" Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.310153 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2"] Mar 17 15:57:42 crc kubenswrapper[4767]: E0317 15:57:42.311828 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36731840-c502-4b79-a9d9-3dcede160c19" containerName="extract" Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.311854 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="36731840-c502-4b79-a9d9-3dcede160c19" containerName="extract" Mar 17 15:57:42 crc kubenswrapper[4767]: E0317 15:57:42.311882 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36731840-c502-4b79-a9d9-3dcede160c19" containerName="util" Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.311890 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="36731840-c502-4b79-a9d9-3dcede160c19" containerName="util" Mar 17 15:57:42 crc kubenswrapper[4767]: E0317 15:57:42.311903 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36731840-c502-4b79-a9d9-3dcede160c19" containerName="pull" Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.311909 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="36731840-c502-4b79-a9d9-3dcede160c19" containerName="pull" Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.312146 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="36731840-c502-4b79-a9d9-3dcede160c19" containerName="extract" Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.313066 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.320071 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-lc7vj" Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.342863 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2"] Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.400487 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pstdr\" (UniqueName: \"kubernetes.io/projected/3ee9406d-d284-417c-83e2-942103e449eb-kube-api-access-pstdr\") pod \"openstack-operator-controller-init-587c5db989-zt4g2\" (UID: \"3ee9406d-d284-417c-83e2-942103e449eb\") " pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.502908 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pstdr\" (UniqueName: \"kubernetes.io/projected/3ee9406d-d284-417c-83e2-942103e449eb-kube-api-access-pstdr\") pod \"openstack-operator-controller-init-587c5db989-zt4g2\" (UID: \"3ee9406d-d284-417c-83e2-942103e449eb\") " pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.527642 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pstdr\" (UniqueName: \"kubernetes.io/projected/3ee9406d-d284-417c-83e2-942103e449eb-kube-api-access-pstdr\") pod \"openstack-operator-controller-init-587c5db989-zt4g2\" (UID: \"3ee9406d-d284-417c-83e2-942103e449eb\") " pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" Mar 17 15:57:42 crc kubenswrapper[4767]: I0317 15:57:42.635360 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" Mar 17 15:57:43 crc kubenswrapper[4767]: I0317 15:57:43.132995 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2"] Mar 17 15:57:43 crc kubenswrapper[4767]: I0317 15:57:43.390480 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" event={"ID":"3ee9406d-d284-417c-83e2-942103e449eb","Type":"ContainerStarted","Data":"8881b46a8785e722d4da4e983195633ad3428d3523935fe8eb07c3d8aecc4342"} Mar 17 15:57:48 crc kubenswrapper[4767]: I0317 15:57:48.476244 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" event={"ID":"3ee9406d-d284-417c-83e2-942103e449eb","Type":"ContainerStarted","Data":"335c58b81be2c4f4da9dae08de9ad825ce914db217cbf1f384ea014203d1ee8c"} Mar 17 15:57:48 crc kubenswrapper[4767]: I0317 15:57:48.476978 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" Mar 17 15:57:48 crc kubenswrapper[4767]: I0317 15:57:48.519855 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" podStartSLOduration=1.455521327 podStartE2EDuration="6.51981773s" podCreationTimestamp="2026-03-17 15:57:42 +0000 UTC" firstStartedPulling="2026-03-17 15:57:43.138306494 +0000 UTC m=+1254.551622541" lastFinishedPulling="2026-03-17 15:57:48.202602897 +0000 UTC m=+1259.615918944" observedRunningTime="2026-03-17 15:57:48.509293246 +0000 UTC m=+1259.922609313" watchObservedRunningTime="2026-03-17 15:57:48.51981773 +0000 UTC m=+1259.933133777" Mar 17 15:58:00 crc kubenswrapper[4767]: I0317 15:58:00.151078 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562718-28mzx"] Mar 17 15:58:00 crc kubenswrapper[4767]: I0317 15:58:00.153955 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562718-28mzx" Mar 17 15:58:00 crc kubenswrapper[4767]: I0317 15:58:00.156725 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 15:58:00 crc kubenswrapper[4767]: I0317 15:58:00.156739 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 15:58:00 crc kubenswrapper[4767]: I0317 15:58:00.156852 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 15:58:00 crc kubenswrapper[4767]: I0317 15:58:00.179713 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562718-28mzx"] Mar 17 15:58:00 crc kubenswrapper[4767]: I0317 15:58:00.352850 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkdvc\" (UniqueName: \"kubernetes.io/projected/a248ba4a-e1d4-4204-b549-7dfff52fe100-kube-api-access-lkdvc\") pod \"auto-csr-approver-29562718-28mzx\" (UID: \"a248ba4a-e1d4-4204-b549-7dfff52fe100\") " pod="openshift-infra/auto-csr-approver-29562718-28mzx" Mar 17 15:58:00 crc kubenswrapper[4767]: I0317 15:58:00.454805 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkdvc\" (UniqueName: \"kubernetes.io/projected/a248ba4a-e1d4-4204-b549-7dfff52fe100-kube-api-access-lkdvc\") pod \"auto-csr-approver-29562718-28mzx\" (UID: \"a248ba4a-e1d4-4204-b549-7dfff52fe100\") " pod="openshift-infra/auto-csr-approver-29562718-28mzx" Mar 17 15:58:00 crc kubenswrapper[4767]: I0317 15:58:00.476582 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkdvc\" (UniqueName: \"kubernetes.io/projected/a248ba4a-e1d4-4204-b549-7dfff52fe100-kube-api-access-lkdvc\") pod \"auto-csr-approver-29562718-28mzx\" (UID: \"a248ba4a-e1d4-4204-b549-7dfff52fe100\") " pod="openshift-infra/auto-csr-approver-29562718-28mzx" Mar 17 15:58:00 crc kubenswrapper[4767]: I0317 15:58:00.494778 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562718-28mzx" Mar 17 15:58:01 crc kubenswrapper[4767]: I0317 15:58:01.055353 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562718-28mzx"] Mar 17 15:58:01 crc kubenswrapper[4767]: I0317 15:58:01.597980 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562718-28mzx" event={"ID":"a248ba4a-e1d4-4204-b549-7dfff52fe100","Type":"ContainerStarted","Data":"499ff0a3abeb88a7f03eabc736a56e06f316936302e40d4f8a8a630a6061c4f2"} Mar 17 15:58:02 crc kubenswrapper[4767]: I0317 15:58:02.644320 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" Mar 17 15:58:03 crc kubenswrapper[4767]: I0317 15:58:03.634779 4767 generic.go:334] "Generic (PLEG): container finished" podID="a248ba4a-e1d4-4204-b549-7dfff52fe100" containerID="20330680ac4fb421b4d9350049f827cc802f164445428f995497f7a928a3b552" exitCode=0 Mar 17 15:58:03 crc kubenswrapper[4767]: I0317 15:58:03.636266 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562718-28mzx" event={"ID":"a248ba4a-e1d4-4204-b549-7dfff52fe100","Type":"ContainerDied","Data":"20330680ac4fb421b4d9350049f827cc802f164445428f995497f7a928a3b552"} Mar 17 15:58:04 crc kubenswrapper[4767]: I0317 15:58:04.996528 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562718-28mzx" Mar 17 15:58:05 crc kubenswrapper[4767]: I0317 15:58:05.189932 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkdvc\" (UniqueName: \"kubernetes.io/projected/a248ba4a-e1d4-4204-b549-7dfff52fe100-kube-api-access-lkdvc\") pod \"a248ba4a-e1d4-4204-b549-7dfff52fe100\" (UID: \"a248ba4a-e1d4-4204-b549-7dfff52fe100\") " Mar 17 15:58:05 crc kubenswrapper[4767]: I0317 15:58:05.200481 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a248ba4a-e1d4-4204-b549-7dfff52fe100-kube-api-access-lkdvc" (OuterVolumeSpecName: "kube-api-access-lkdvc") pod "a248ba4a-e1d4-4204-b549-7dfff52fe100" (UID: "a248ba4a-e1d4-4204-b549-7dfff52fe100"). InnerVolumeSpecName "kube-api-access-lkdvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 15:58:05 crc kubenswrapper[4767]: I0317 15:58:05.292110 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkdvc\" (UniqueName: \"kubernetes.io/projected/a248ba4a-e1d4-4204-b549-7dfff52fe100-kube-api-access-lkdvc\") on node \"crc\" DevicePath \"\"" Mar 17 15:58:05 crc kubenswrapper[4767]: I0317 15:58:05.658626 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562718-28mzx" event={"ID":"a248ba4a-e1d4-4204-b549-7dfff52fe100","Type":"ContainerDied","Data":"499ff0a3abeb88a7f03eabc736a56e06f316936302e40d4f8a8a630a6061c4f2"} Mar 17 15:58:05 crc kubenswrapper[4767]: I0317 15:58:05.658698 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="499ff0a3abeb88a7f03eabc736a56e06f316936302e40d4f8a8a630a6061c4f2" Mar 17 15:58:05 crc kubenswrapper[4767]: I0317 15:58:05.658697 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562718-28mzx" Mar 17 15:58:06 crc kubenswrapper[4767]: I0317 15:58:06.066323 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562712-r589g"] Mar 17 15:58:06 crc kubenswrapper[4767]: I0317 15:58:06.073991 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562712-r589g"] Mar 17 15:58:07 crc kubenswrapper[4767]: I0317 15:58:07.370425 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4ed68ee-0956-4d37-9d3b-f1904ff19451" path="/var/lib/kubelet/pods/f4ed68ee-0956-4d37-9d3b-f1904ff19451/volumes" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.616666 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc"] Mar 17 15:58:32 crc kubenswrapper[4767]: E0317 15:58:32.618328 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a248ba4a-e1d4-4204-b549-7dfff52fe100" containerName="oc" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.618354 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a248ba4a-e1d4-4204-b549-7dfff52fe100" containerName="oc" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.618610 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a248ba4a-e1d4-4204-b549-7dfff52fe100" containerName="oc" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.619604 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.622525 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-k4gp4" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.634217 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995"] Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.635786 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.644609 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-mxvqg" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.648345 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5"] Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.649961 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.653981 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-qmdh2" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.658241 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc"] Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.682271 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995"] Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.732229 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw"] Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.733563 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.744026 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmq67\" (UniqueName: \"kubernetes.io/projected/53aea289-5a28-438b-8d28-242d836351f0-kube-api-access-mmq67\") pod \"barbican-operator-controller-manager-59bc569d95-rvkxc\" (UID: \"53aea289-5a28-438b-8d28-242d836351f0\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.744109 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh6sb\" (UniqueName: \"kubernetes.io/projected/b6dc3b02-8dc0-46ac-b78e-1d7d3ffb3f24-kube-api-access-sh6sb\") pod \"cinder-operator-controller-manager-8d58dc466-k8995\" (UID: \"b6dc3b02-8dc0-46ac-b78e-1d7d3ffb3f24\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.753715 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-5jwqk" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.798382 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw"] Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.843351 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n"] Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.845426 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.846218 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xcrq\" (UniqueName: \"kubernetes.io/projected/16e4d9a1-285b-4221-8a99-55d515bc3356-kube-api-access-5xcrq\") pod \"glance-operator-controller-manager-79df6bcc97-p2skw\" (UID: \"16e4d9a1-285b-4221-8a99-55d515bc3356\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.846432 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjgx4\" (UniqueName: \"kubernetes.io/projected/a67a8ddc-0801-4530-bf82-9d4f4d0389a8-kube-api-access-hjgx4\") pod \"designate-operator-controller-manager-588d4d986b-q7zd5\" (UID: \"a67a8ddc-0801-4530-bf82-9d4f4d0389a8\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.849460 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmq67\" (UniqueName: \"kubernetes.io/projected/53aea289-5a28-438b-8d28-242d836351f0-kube-api-access-mmq67\") pod \"barbican-operator-controller-manager-59bc569d95-rvkxc\" (UID: \"53aea289-5a28-438b-8d28-242d836351f0\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.857650 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh6sb\" (UniqueName: \"kubernetes.io/projected/b6dc3b02-8dc0-46ac-b78e-1d7d3ffb3f24-kube-api-access-sh6sb\") pod \"cinder-operator-controller-manager-8d58dc466-k8995\" (UID: \"b6dc3b02-8dc0-46ac-b78e-1d7d3ffb3f24\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.854043 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-kdmm2" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.852325 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c"] Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.863881 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.870105 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5"] Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.877346 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-rfxxp" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.885786 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n"] Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.899769 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmq67\" (UniqueName: \"kubernetes.io/projected/53aea289-5a28-438b-8d28-242d836351f0-kube-api-access-mmq67\") pod \"barbican-operator-controller-manager-59bc569d95-rvkxc\" (UID: \"53aea289-5a28-438b-8d28-242d836351f0\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.917509 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh6sb\" (UniqueName: \"kubernetes.io/projected/b6dc3b02-8dc0-46ac-b78e-1d7d3ffb3f24-kube-api-access-sh6sb\") pod \"cinder-operator-controller-manager-8d58dc466-k8995\" (UID: \"b6dc3b02-8dc0-46ac-b78e-1d7d3ffb3f24\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.935306 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2"] Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.938011 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.955709 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.960648 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert\") pod \"infra-operator-controller-manager-7b9c774f96-kftt2\" (UID: \"95d5d3e8-dc72-414f-afe4-b68b757a39d4\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.960779 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjgx4\" (UniqueName: \"kubernetes.io/projected/a67a8ddc-0801-4530-bf82-9d4f4d0389a8-kube-api-access-hjgx4\") pod \"designate-operator-controller-manager-588d4d986b-q7zd5\" (UID: \"a67a8ddc-0801-4530-bf82-9d4f4d0389a8\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.960945 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68x6g\" (UniqueName: \"kubernetes.io/projected/0163654c-d57e-4b14-aba0-f76dbaff1114-kube-api-access-68x6g\") pod \"heat-operator-controller-manager-67dd5f86f5-drx6n\" (UID: \"0163654c-d57e-4b14-aba0-f76dbaff1114\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.961014 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj5mp\" (UniqueName: \"kubernetes.io/projected/95d5d3e8-dc72-414f-afe4-b68b757a39d4-kube-api-access-tj5mp\") pod \"infra-operator-controller-manager-7b9c774f96-kftt2\" (UID: \"95d5d3e8-dc72-414f-afe4-b68b757a39d4\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.961068 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l84l2\" (UniqueName: \"kubernetes.io/projected/fa1e1603-adc4-4ade-aca8-e59df8c350b2-kube-api-access-l84l2\") pod \"horizon-operator-controller-manager-8464cc45fb-wj49c\" (UID: \"fa1e1603-adc4-4ade-aca8-e59df8c350b2\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.961158 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xcrq\" (UniqueName: \"kubernetes.io/projected/16e4d9a1-285b-4221-8a99-55d515bc3356-kube-api-access-5xcrq\") pod \"glance-operator-controller-manager-79df6bcc97-p2skw\" (UID: \"16e4d9a1-285b-4221-8a99-55d515bc3356\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" Mar 17 15:58:32 crc kubenswrapper[4767]: I0317 15:58:32.975901 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:32.981396 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c"] Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.072681 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert\") pod \"infra-operator-controller-manager-7b9c774f96-kftt2\" (UID: \"95d5d3e8-dc72-414f-afe4-b68b757a39d4\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.072841 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68x6g\" (UniqueName: \"kubernetes.io/projected/0163654c-d57e-4b14-aba0-f76dbaff1114-kube-api-access-68x6g\") pod \"heat-operator-controller-manager-67dd5f86f5-drx6n\" (UID: \"0163654c-d57e-4b14-aba0-f76dbaff1114\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.072877 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj5mp\" (UniqueName: \"kubernetes.io/projected/95d5d3e8-dc72-414f-afe4-b68b757a39d4-kube-api-access-tj5mp\") pod \"infra-operator-controller-manager-7b9c774f96-kftt2\" (UID: \"95d5d3e8-dc72-414f-afe4-b68b757a39d4\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.072921 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l84l2\" (UniqueName: \"kubernetes.io/projected/fa1e1603-adc4-4ade-aca8-e59df8c350b2-kube-api-access-l84l2\") pod \"horizon-operator-controller-manager-8464cc45fb-wj49c\" (UID: \"fa1e1603-adc4-4ade-aca8-e59df8c350b2\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.105631 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.105931 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-x82p6" Mar 17 15:58:33 crc kubenswrapper[4767]: E0317 15:58:33.117373 4767 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 17 15:58:33 crc kubenswrapper[4767]: E0317 15:58:33.117515 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert podName:95d5d3e8-dc72-414f-afe4-b68b757a39d4 nodeName:}" failed. No retries permitted until 2026-03-17 15:58:33.617478533 +0000 UTC m=+1305.030794570 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert") pod "infra-operator-controller-manager-7b9c774f96-kftt2" (UID: "95d5d3e8-dc72-414f-afe4-b68b757a39d4") : secret "infra-operator-webhook-server-cert" not found Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.140034 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xcrq\" (UniqueName: \"kubernetes.io/projected/16e4d9a1-285b-4221-8a99-55d515bc3356-kube-api-access-5xcrq\") pod \"glance-operator-controller-manager-79df6bcc97-p2skw\" (UID: \"16e4d9a1-285b-4221-8a99-55d515bc3356\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.145282 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj"] Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.147065 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.159996 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-5l2jc" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.172896 4767 scope.go:117] "RemoveContainer" containerID="63135d6512e4bb5413b9841549fb5ec1a52f9e8411d5691cb6707e9465c12312" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.188653 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjgx4\" (UniqueName: \"kubernetes.io/projected/a67a8ddc-0801-4530-bf82-9d4f4d0389a8-kube-api-access-hjgx4\") pod \"designate-operator-controller-manager-588d4d986b-q7zd5\" (UID: \"a67a8ddc-0801-4530-bf82-9d4f4d0389a8\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.190685 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68x6g\" (UniqueName: \"kubernetes.io/projected/0163654c-d57e-4b14-aba0-f76dbaff1114-kube-api-access-68x6g\") pod \"heat-operator-controller-manager-67dd5f86f5-drx6n\" (UID: \"0163654c-d57e-4b14-aba0-f76dbaff1114\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.221093 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l84l2\" (UniqueName: \"kubernetes.io/projected/fa1e1603-adc4-4ade-aca8-e59df8c350b2-kube-api-access-l84l2\") pod \"horizon-operator-controller-manager-8464cc45fb-wj49c\" (UID: \"fa1e1603-adc4-4ade-aca8-e59df8c350b2\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.221891 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj5mp\" (UniqueName: \"kubernetes.io/projected/95d5d3e8-dc72-414f-afe4-b68b757a39d4-kube-api-access-tj5mp\") pod \"infra-operator-controller-manager-7b9c774f96-kftt2\" (UID: \"95d5d3e8-dc72-414f-afe4-b68b757a39d4\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.223233 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj"] Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.295195 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc8m5\" (UniqueName: \"kubernetes.io/projected/904d9f3b-95f0-4e57-8d04-c2026227a4a6-kube-api-access-fc8m5\") pod \"ironic-operator-controller-manager-6f787dddc9-24wkj\" (UID: \"904d9f3b-95f0-4e57-8d04-c2026227a4a6\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.299530 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2"] Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.299766 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.338319 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv"] Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.339969 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.353929 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-6lhbn" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.378752 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.402398 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc8m5\" (UniqueName: \"kubernetes.io/projected/904d9f3b-95f0-4e57-8d04-c2026227a4a6-kube-api-access-fc8m5\") pod \"ironic-operator-controller-manager-6f787dddc9-24wkj\" (UID: \"904d9f3b-95f0-4e57-8d04-c2026227a4a6\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.399160 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv"] Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.448306 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc8m5\" (UniqueName: \"kubernetes.io/projected/904d9f3b-95f0-4e57-8d04-c2026227a4a6-kube-api-access-fc8m5\") pod \"ironic-operator-controller-manager-6f787dddc9-24wkj\" (UID: \"904d9f3b-95f0-4e57-8d04-c2026227a4a6\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.472762 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.476549 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-xfv84"] Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.478099 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.489977 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-5nl6z" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.501943 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.514372 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfbgg\" (UniqueName: \"kubernetes.io/projected/a7b5a1f5-da92-46dd-a5b0-5088e75346af-kube-api-access-zfbgg\") pod \"keystone-operator-controller-manager-768b96df4c-x6vcv\" (UID: \"a7b5a1f5-da92-46dd-a5b0-5088e75346af\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.516273 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-xfv84"] Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.575227 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq"] Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.576913 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.579628 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.591303 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-jb8nz" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.617049 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-btzsh"] Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.618910 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.636562 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtbbb\" (UniqueName: \"kubernetes.io/projected/43b467d2-9860-4feb-a656-65827836c23e-kube-api-access-qtbbb\") pod \"mariadb-operator-controller-manager-67ccfc9778-ks6jq\" (UID: \"43b467d2-9860-4feb-a656-65827836c23e\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.669623 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfbgg\" (UniqueName: \"kubernetes.io/projected/a7b5a1f5-da92-46dd-a5b0-5088e75346af-kube-api-access-zfbgg\") pod \"keystone-operator-controller-manager-768b96df4c-x6vcv\" (UID: \"a7b5a1f5-da92-46dd-a5b0-5088e75346af\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.669751 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert\") pod \"infra-operator-controller-manager-7b9c774f96-kftt2\" (UID: \"95d5d3e8-dc72-414f-afe4-b68b757a39d4\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.669837 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh9d5\" (UniqueName: \"kubernetes.io/projected/81201888-f7c3-4605-86a7-85f3edbca3a6-kube-api-access-zh9d5\") pod \"manila-operator-controller-manager-55f864c847-xfv84\" (UID: \"81201888-f7c3-4605-86a7-85f3edbca3a6\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" Mar 17 15:58:33 crc kubenswrapper[4767]: E0317 15:58:33.670604 4767 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 17 15:58:33 crc kubenswrapper[4767]: E0317 15:58:33.670674 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert podName:95d5d3e8-dc72-414f-afe4-b68b757a39d4 nodeName:}" failed. No retries permitted until 2026-03-17 15:58:34.670652342 +0000 UTC m=+1306.083968389 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert") pod "infra-operator-controller-manager-7b9c774f96-kftt2" (UID: "95d5d3e8-dc72-414f-afe4-b68b757a39d4") : secret "infra-operator-webhook-server-cert" not found Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.689949 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-jmk86" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.766466 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfbgg\" (UniqueName: \"kubernetes.io/projected/a7b5a1f5-da92-46dd-a5b0-5088e75346af-kube-api-access-zfbgg\") pod \"keystone-operator-controller-manager-768b96df4c-x6vcv\" (UID: \"a7b5a1f5-da92-46dd-a5b0-5088e75346af\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.768646 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq"] Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.934021 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtbbb\" (UniqueName: \"kubernetes.io/projected/43b467d2-9860-4feb-a656-65827836c23e-kube-api-access-qtbbb\") pod \"mariadb-operator-controller-manager-67ccfc9778-ks6jq\" (UID: \"43b467d2-9860-4feb-a656-65827836c23e\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.948872 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qmtn\" (UniqueName: \"kubernetes.io/projected/8bfc70f3-ead2-43cc-9e90-7df32804d9ac-kube-api-access-6qmtn\") pod \"neutron-operator-controller-manager-767865f676-btzsh\" (UID: \"8bfc70f3-ead2-43cc-9e90-7df32804d9ac\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" Mar 17 15:58:33 crc kubenswrapper[4767]: I0317 15:58:33.949145 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh9d5\" (UniqueName: \"kubernetes.io/projected/81201888-f7c3-4605-86a7-85f3edbca3a6-kube-api-access-zh9d5\") pod \"manila-operator-controller-manager-55f864c847-xfv84\" (UID: \"81201888-f7c3-4605-86a7-85f3edbca3a6\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.073213 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.074103 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.075986 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.089354 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-p6whc" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.107073 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtbbb\" (UniqueName: \"kubernetes.io/projected/43b467d2-9860-4feb-a656-65827836c23e-kube-api-access-qtbbb\") pod \"mariadb-operator-controller-manager-67ccfc9778-ks6jq\" (UID: \"43b467d2-9860-4feb-a656-65827836c23e\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.148057 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh9d5\" (UniqueName: \"kubernetes.io/projected/81201888-f7c3-4605-86a7-85f3edbca3a6-kube-api-access-zh9d5\") pod \"manila-operator-controller-manager-55f864c847-xfv84\" (UID: \"81201888-f7c3-4605-86a7-85f3edbca3a6\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.149557 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqqn6\" (UniqueName: \"kubernetes.io/projected/921c5e77-9858-4177-99a0-8cd3a4420d7d-kube-api-access-gqqn6\") pod \"nova-operator-controller-manager-5d488d59fb-r8z96\" (UID: \"921c5e77-9858-4177-99a0-8cd3a4420d7d\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.149617 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qmtn\" (UniqueName: \"kubernetes.io/projected/8bfc70f3-ead2-43cc-9e90-7df32804d9ac-kube-api-access-6qmtn\") pod \"neutron-operator-controller-manager-767865f676-btzsh\" (UID: \"8bfc70f3-ead2-43cc-9e90-7df32804d9ac\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.169341 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-btzsh"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.181941 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qmtn\" (UniqueName: \"kubernetes.io/projected/8bfc70f3-ead2-43cc-9e90-7df32804d9ac-kube-api-access-6qmtn\") pod \"neutron-operator-controller-manager-767865f676-btzsh\" (UID: \"8bfc70f3-ead2-43cc-9e90-7df32804d9ac\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.195785 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.204556 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.253735 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqqn6\" (UniqueName: \"kubernetes.io/projected/921c5e77-9858-4177-99a0-8cd3a4420d7d-kube-api-access-gqqn6\") pod \"nova-operator-controller-manager-5d488d59fb-r8z96\" (UID: \"921c5e77-9858-4177-99a0-8cd3a4420d7d\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.281997 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.282041 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-44q8c"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.283317 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.284191 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-44q8c"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.284209 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.285226 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.285881 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.287685 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.287875 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.290758 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.298134 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.308611 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-xp95p" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.310765 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.310940 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-xkm4z" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.311652 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-ggtbx" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.325077 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-hwtx7" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.325361 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.326870 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.334340 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-5pvl9" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.335071 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.341073 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqqn6\" (UniqueName: \"kubernetes.io/projected/921c5e77-9858-4177-99a0-8cd3a4420d7d-kube-api-access-gqqn6\") pod \"nova-operator-controller-manager-5d488d59fb-r8z96\" (UID: \"921c5e77-9858-4177-99a0-8cd3a4420d7d\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.343316 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.344994 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.348404 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-9j9vm" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.361771 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.363188 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4gcn\" (UniqueName: \"kubernetes.io/projected/70fbe8aa-8647-4fe8-914c-0c05399cf46d-kube-api-access-z4gcn\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-pjldm\" (UID: \"70fbe8aa-8647-4fe8-914c-0c05399cf46d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.363540 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlc7d\" (UniqueName: \"kubernetes.io/projected/bd2dbd72-69bf-40c1-b591-be3782c33465-kube-api-access-zlc7d\") pod \"placement-operator-controller-manager-5784578c99-2k2dh\" (UID: \"bd2dbd72-69bf-40c1-b591-be3782c33465\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.363689 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4xfl\" (UniqueName: \"kubernetes.io/projected/59ba3b25-1e95-41ad-921c-9ee4ec5e2c43-kube-api-access-j4xfl\") pod \"octavia-operator-controller-manager-5b9f45d989-xms5z\" (UID: \"59ba3b25-1e95-41ad-921c-9ee4ec5e2c43\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.363783 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-pjldm\" (UID: \"70fbe8aa-8647-4fe8-914c-0c05399cf46d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.363834 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2dsb\" (UniqueName: \"kubernetes.io/projected/a833dc16-3a29-4129-b592-732d71818bc4-kube-api-access-h2dsb\") pod \"ovn-operator-controller-manager-884679f54-44q8c\" (UID: \"a833dc16-3a29-4129-b592-732d71818bc4\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.363935 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.374298 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.375448 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.383865 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-n4ps2" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.402998 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.421552 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.424336 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.458747 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.460455 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.469434 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4wp7\" (UniqueName: \"kubernetes.io/projected/ad853953-ec48-40fc-8787-b2b838c955e9-kube-api-access-b4wp7\") pod \"swift-operator-controller-manager-c674c5965-jx2k5\" (UID: \"ad853953-ec48-40fc-8787-b2b838c955e9\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.469515 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4psx\" (UniqueName: \"kubernetes.io/projected/e59ff072-51c2-4995-9f61-709f12a1393a-kube-api-access-b4psx\") pod \"test-operator-controller-manager-5c5cb9c4d7-b8f8r\" (UID: \"e59ff072-51c2-4995-9f61-709f12a1393a\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.469585 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4xfl\" (UniqueName: \"kubernetes.io/projected/59ba3b25-1e95-41ad-921c-9ee4ec5e2c43-kube-api-access-j4xfl\") pod \"octavia-operator-controller-manager-5b9f45d989-xms5z\" (UID: \"59ba3b25-1e95-41ad-921c-9ee4ec5e2c43\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.469643 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvf7z\" (UniqueName: \"kubernetes.io/projected/7381ff68-5e5d-4281-9924-8495ece760f4-kube-api-access-mvf7z\") pod \"telemetry-operator-controller-manager-766dd46678-n9j9q\" (UID: \"7381ff68-5e5d-4281-9924-8495ece760f4\") " pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.469671 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-pjldm\" (UID: \"70fbe8aa-8647-4fe8-914c-0c05399cf46d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.469706 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2dsb\" (UniqueName: \"kubernetes.io/projected/a833dc16-3a29-4129-b592-732d71818bc4-kube-api-access-h2dsb\") pod \"ovn-operator-controller-manager-884679f54-44q8c\" (UID: \"a833dc16-3a29-4129-b592-732d71818bc4\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.469762 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4gcn\" (UniqueName: \"kubernetes.io/projected/70fbe8aa-8647-4fe8-914c-0c05399cf46d-kube-api-access-z4gcn\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-pjldm\" (UID: \"70fbe8aa-8647-4fe8-914c-0c05399cf46d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.469841 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlc7d\" (UniqueName: \"kubernetes.io/projected/bd2dbd72-69bf-40c1-b591-be3782c33465-kube-api-access-zlc7d\") pod \"placement-operator-controller-manager-5784578c99-2k2dh\" (UID: \"bd2dbd72-69bf-40c1-b591-be3782c33465\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" Mar 17 15:58:34 crc kubenswrapper[4767]: E0317 15:58:34.473736 4767 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 15:58:34 crc kubenswrapper[4767]: E0317 15:58:34.473871 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert podName:70fbe8aa-8647-4fe8-914c-0c05399cf46d nodeName:}" failed. No retries permitted until 2026-03-17 15:58:34.973838961 +0000 UTC m=+1306.387155008 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-pjldm" (UID: "70fbe8aa-8647-4fe8-914c-0c05399cf46d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.474649 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-7vc25" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.503648 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4gcn\" (UniqueName: \"kubernetes.io/projected/70fbe8aa-8647-4fe8-914c-0c05399cf46d-kube-api-access-z4gcn\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-pjldm\" (UID: \"70fbe8aa-8647-4fe8-914c-0c05399cf46d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.510208 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2dsb\" (UniqueName: \"kubernetes.io/projected/a833dc16-3a29-4129-b592-732d71818bc4-kube-api-access-h2dsb\") pod \"ovn-operator-controller-manager-884679f54-44q8c\" (UID: \"a833dc16-3a29-4129-b592-732d71818bc4\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.511849 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlc7d\" (UniqueName: \"kubernetes.io/projected/bd2dbd72-69bf-40c1-b591-be3782c33465-kube-api-access-zlc7d\") pod \"placement-operator-controller-manager-5784578c99-2k2dh\" (UID: \"bd2dbd72-69bf-40c1-b591-be3782c33465\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.517078 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4xfl\" (UniqueName: \"kubernetes.io/projected/59ba3b25-1e95-41ad-921c-9ee4ec5e2c43-kube-api-access-j4xfl\") pod \"octavia-operator-controller-manager-5b9f45d989-xms5z\" (UID: \"59ba3b25-1e95-41ad-921c-9ee4ec5e2c43\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.519490 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.551969 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.560134 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.561695 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.568900 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.576817 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.577020 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.577243 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-f4m4f" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.583612 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4wp7\" (UniqueName: \"kubernetes.io/projected/ad853953-ec48-40fc-8787-b2b838c955e9-kube-api-access-b4wp7\") pod \"swift-operator-controller-manager-c674c5965-jx2k5\" (UID: \"ad853953-ec48-40fc-8787-b2b838c955e9\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.583756 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4psx\" (UniqueName: \"kubernetes.io/projected/e59ff072-51c2-4995-9f61-709f12a1393a-kube-api-access-b4psx\") pod \"test-operator-controller-manager-5c5cb9c4d7-b8f8r\" (UID: \"e59ff072-51c2-4995-9f61-709f12a1393a\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.583974 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvf7z\" (UniqueName: \"kubernetes.io/projected/7381ff68-5e5d-4281-9924-8495ece760f4-kube-api-access-mvf7z\") pod \"telemetry-operator-controller-manager-766dd46678-n9j9q\" (UID: \"7381ff68-5e5d-4281-9924-8495ece760f4\") " pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.584030 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfd8d\" (UniqueName: \"kubernetes.io/projected/faf77dfd-c76e-4abd-a8c8-b6211cf070a4-kube-api-access-mfd8d\") pod \"watcher-operator-controller-manager-6c4d75f7f9-22vp8\" (UID: \"faf77dfd-c76e-4abd-a8c8-b6211cf070a4\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.622477 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.624576 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.626655 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4wp7\" (UniqueName: \"kubernetes.io/projected/ad853953-ec48-40fc-8787-b2b838c955e9-kube-api-access-b4wp7\") pod \"swift-operator-controller-manager-c674c5965-jx2k5\" (UID: \"ad853953-ec48-40fc-8787-b2b838c955e9\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.628475 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4psx\" (UniqueName: \"kubernetes.io/projected/e59ff072-51c2-4995-9f61-709f12a1393a-kube-api-access-b4psx\") pod \"test-operator-controller-manager-5c5cb9c4d7-b8f8r\" (UID: \"e59ff072-51c2-4995-9f61-709f12a1393a\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.631390 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-hmcxt" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.632255 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvf7z\" (UniqueName: \"kubernetes.io/projected/7381ff68-5e5d-4281-9924-8495ece760f4-kube-api-access-mvf7z\") pod \"telemetry-operator-controller-manager-766dd46678-n9j9q\" (UID: \"7381ff68-5e5d-4281-9924-8495ece760f4\") " pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.632261 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.643346 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.676635 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.687716 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.690541 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert\") pod \"infra-operator-controller-manager-7b9c774f96-kftt2\" (UID: \"95d5d3e8-dc72-414f-afe4-b68b757a39d4\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.691000 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvps9\" (UniqueName: \"kubernetes.io/projected/713154c5-7e16-498f-9612-1c0afbf362bb-kube-api-access-fvps9\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.691122 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfd8d\" (UniqueName: \"kubernetes.io/projected/faf77dfd-c76e-4abd-a8c8-b6211cf070a4-kube-api-access-mfd8d\") pod \"watcher-operator-controller-manager-6c4d75f7f9-22vp8\" (UID: \"faf77dfd-c76e-4abd-a8c8-b6211cf070a4\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.691599 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:34 crc kubenswrapper[4767]: E0317 15:58:34.691934 4767 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 17 15:58:34 crc kubenswrapper[4767]: E0317 15:58:34.692041 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert podName:95d5d3e8-dc72-414f-afe4-b68b757a39d4 nodeName:}" failed. No retries permitted until 2026-03-17 15:58:36.692012826 +0000 UTC m=+1308.105328873 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert") pod "infra-operator-controller-manager-7b9c774f96-kftt2" (UID: "95d5d3e8-dc72-414f-afe4-b68b757a39d4") : secret "infra-operator-webhook-server-cert" not found Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.714112 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.756232 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfd8d\" (UniqueName: \"kubernetes.io/projected/faf77dfd-c76e-4abd-a8c8-b6211cf070a4-kube-api-access-mfd8d\") pod \"watcher-operator-controller-manager-6c4d75f7f9-22vp8\" (UID: \"faf77dfd-c76e-4abd-a8c8-b6211cf070a4\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" Mar 17 15:58:34 crc kubenswrapper[4767]: E0317 15:58:34.794615 4767 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 17 15:58:34 crc kubenswrapper[4767]: E0317 15:58:34.794756 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs podName:713154c5-7e16-498f-9612-1c0afbf362bb nodeName:}" failed. No retries permitted until 2026-03-17 15:58:35.294711348 +0000 UTC m=+1306.708027395 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs") pod "openstack-operator-controller-manager-d7dccc75b-zlss6" (UID: "713154c5-7e16-498f-9612-1c0afbf362bb") : secret "webhook-server-cert" not found Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.794389 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.794831 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl9fj\" (UniqueName: \"kubernetes.io/projected/248e2576-adbb-40bb-ade4-8c426f2a8a48-kube-api-access-jl9fj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tvpkb\" (UID: \"248e2576-adbb-40bb-ade4-8c426f2a8a48\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.794999 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvps9\" (UniqueName: \"kubernetes.io/projected/713154c5-7e16-498f-9612-1c0afbf362bb-kube-api-access-fvps9\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.795100 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:34 crc kubenswrapper[4767]: E0317 15:58:34.795817 4767 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 17 15:58:34 crc kubenswrapper[4767]: E0317 15:58:34.795919 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs podName:713154c5-7e16-498f-9612-1c0afbf362bb nodeName:}" failed. No retries permitted until 2026-03-17 15:58:35.295892549 +0000 UTC m=+1306.709208596 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs") pod "openstack-operator-controller-manager-d7dccc75b-zlss6" (UID: "713154c5-7e16-498f-9612-1c0afbf362bb") : secret "metrics-server-cert" not found Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.806792 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.837079 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvps9\" (UniqueName: \"kubernetes.io/projected/713154c5-7e16-498f-9612-1c0afbf362bb-kube-api-access-fvps9\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.850952 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.851361 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.862789 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.892709 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc"] Mar 17 15:58:34 crc kubenswrapper[4767]: I0317 15:58:34.904675 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl9fj\" (UniqueName: \"kubernetes.io/projected/248e2576-adbb-40bb-ade4-8c426f2a8a48-kube-api-access-jl9fj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tvpkb\" (UID: \"248e2576-adbb-40bb-ade4-8c426f2a8a48\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb" Mar 17 15:58:35 crc kubenswrapper[4767]: I0317 15:58:35.172315 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-pjldm\" (UID: \"70fbe8aa-8647-4fe8-914c-0c05399cf46d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:35 crc kubenswrapper[4767]: E0317 15:58:35.173110 4767 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 15:58:35 crc kubenswrapper[4767]: E0317 15:58:35.173187 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert podName:70fbe8aa-8647-4fe8-914c-0c05399cf46d nodeName:}" failed. No retries permitted until 2026-03-17 15:58:36.173159417 +0000 UTC m=+1307.586475464 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-pjldm" (UID: "70fbe8aa-8647-4fe8-914c-0c05399cf46d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 15:58:35 crc kubenswrapper[4767]: I0317 15:58:35.231562 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl9fj\" (UniqueName: \"kubernetes.io/projected/248e2576-adbb-40bb-ade4-8c426f2a8a48-kube-api-access-jl9fj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tvpkb\" (UID: \"248e2576-adbb-40bb-ade4-8c426f2a8a48\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb" Mar 17 15:58:35 crc kubenswrapper[4767]: I0317 15:58:35.384321 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:35 crc kubenswrapper[4767]: I0317 15:58:35.384471 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:35 crc kubenswrapper[4767]: E0317 15:58:35.384758 4767 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 17 15:58:35 crc kubenswrapper[4767]: E0317 15:58:35.384841 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs podName:713154c5-7e16-498f-9612-1c0afbf362bb nodeName:}" failed. No retries permitted until 2026-03-17 15:58:36.384818714 +0000 UTC m=+1307.798134761 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs") pod "openstack-operator-controller-manager-d7dccc75b-zlss6" (UID: "713154c5-7e16-498f-9612-1c0afbf362bb") : secret "webhook-server-cert" not found Mar 17 15:58:35 crc kubenswrapper[4767]: E0317 15:58:35.384847 4767 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 17 15:58:35 crc kubenswrapper[4767]: E0317 15:58:35.384887 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs podName:713154c5-7e16-498f-9612-1c0afbf362bb nodeName:}" failed. No retries permitted until 2026-03-17 15:58:36.384875585 +0000 UTC m=+1307.798191642 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs") pod "openstack-operator-controller-manager-d7dccc75b-zlss6" (UID: "713154c5-7e16-498f-9612-1c0afbf362bb") : secret "metrics-server-cert" not found Mar 17 15:58:35 crc kubenswrapper[4767]: I0317 15:58:35.439197 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" Mar 17 15:58:35 crc kubenswrapper[4767]: I0317 15:58:35.448552 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" event={"ID":"b6dc3b02-8dc0-46ac-b78e-1d7d3ffb3f24","Type":"ContainerStarted","Data":"a4881b692a412ab58aeefe53f42ec1223aba51b8d1c9d5c2b0fa446938c35f14"} Mar 17 15:58:35 crc kubenswrapper[4767]: I0317 15:58:35.456675 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" event={"ID":"53aea289-5a28-438b-8d28-242d836351f0","Type":"ContainerStarted","Data":"eea66143d1323400761028ffd7f7bfcd633516da59d21371a71d9690153e98f3"} Mar 17 15:58:35 crc kubenswrapper[4767]: I0317 15:58:35.531373 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb" Mar 17 15:58:36 crc kubenswrapper[4767]: I0317 15:58:36.204462 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-pjldm\" (UID: \"70fbe8aa-8647-4fe8-914c-0c05399cf46d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:36 crc kubenswrapper[4767]: E0317 15:58:36.204812 4767 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 15:58:36 crc kubenswrapper[4767]: E0317 15:58:36.205385 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert podName:70fbe8aa-8647-4fe8-914c-0c05399cf46d nodeName:}" failed. No retries permitted until 2026-03-17 15:58:38.205353162 +0000 UTC m=+1309.618669209 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-pjldm" (UID: "70fbe8aa-8647-4fe8-914c-0c05399cf46d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 15:58:36 crc kubenswrapper[4767]: I0317 15:58:36.408301 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:36 crc kubenswrapper[4767]: I0317 15:58:36.408466 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:36 crc kubenswrapper[4767]: E0317 15:58:36.408688 4767 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 17 15:58:36 crc kubenswrapper[4767]: E0317 15:58:36.408763 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs podName:713154c5-7e16-498f-9612-1c0afbf362bb nodeName:}" failed. No retries permitted until 2026-03-17 15:58:38.408740304 +0000 UTC m=+1309.822056351 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs") pod "openstack-operator-controller-manager-d7dccc75b-zlss6" (UID: "713154c5-7e16-498f-9612-1c0afbf362bb") : secret "metrics-server-cert" not found Mar 17 15:58:36 crc kubenswrapper[4767]: E0317 15:58:36.409239 4767 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 17 15:58:36 crc kubenswrapper[4767]: E0317 15:58:36.409265 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs podName:713154c5-7e16-498f-9612-1c0afbf362bb nodeName:}" failed. No retries permitted until 2026-03-17 15:58:38.409256717 +0000 UTC m=+1309.822572764 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs") pod "openstack-operator-controller-manager-d7dccc75b-zlss6" (UID: "713154c5-7e16-498f-9612-1c0afbf362bb") : secret "webhook-server-cert" not found Mar 17 15:58:36 crc kubenswrapper[4767]: I0317 15:58:36.530601 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n"] Mar 17 15:58:36 crc kubenswrapper[4767]: I0317 15:58:36.568124 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv"] Mar 17 15:58:36 crc kubenswrapper[4767]: I0317 15:58:36.586647 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw"] Mar 17 15:58:36 crc kubenswrapper[4767]: I0317 15:58:36.611966 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5"] Mar 17 15:58:36 crc kubenswrapper[4767]: I0317 15:58:36.621218 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj"] Mar 17 15:58:36 crc kubenswrapper[4767]: W0317 15:58:36.644980 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda67a8ddc_0801_4530_bf82_9d4f4d0389a8.slice/crio-081b1095b3c6a82b60bb08e58ef4958d6585ec9d90dc392629bc9104ad068367 WatchSource:0}: Error finding container 081b1095b3c6a82b60bb08e58ef4958d6585ec9d90dc392629bc9104ad068367: Status 404 returned error can't find the container with id 081b1095b3c6a82b60bb08e58ef4958d6585ec9d90dc392629bc9104ad068367 Mar 17 15:58:36 crc kubenswrapper[4767]: I0317 15:58:36.717842 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert\") pod \"infra-operator-controller-manager-7b9c774f96-kftt2\" (UID: \"95d5d3e8-dc72-414f-afe4-b68b757a39d4\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:36 crc kubenswrapper[4767]: E0317 15:58:36.723258 4767 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 17 15:58:36 crc kubenswrapper[4767]: E0317 15:58:36.723429 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert podName:95d5d3e8-dc72-414f-afe4-b68b757a39d4 nodeName:}" failed. No retries permitted until 2026-03-17 15:58:40.72338953 +0000 UTC m=+1312.136705667 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert") pod "infra-operator-controller-manager-7b9c774f96-kftt2" (UID: "95d5d3e8-dc72-414f-afe4-b68b757a39d4") : secret "infra-operator-webhook-server-cert" not found Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.011602 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh"] Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.045048 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z"] Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.053207 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq"] Mar 17 15:58:37 crc kubenswrapper[4767]: W0317 15:58:37.087719 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59ba3b25_1e95_41ad_921c_9ee4ec5e2c43.slice/crio-1a748f50ca8fcb8f1ac7b4cdc59d946f07b702a86564a5bae9c0fee0d212d688 WatchSource:0}: Error finding container 1a748f50ca8fcb8f1ac7b4cdc59d946f07b702a86564a5bae9c0fee0d212d688: Status 404 returned error can't find the container with id 1a748f50ca8fcb8f1ac7b4cdc59d946f07b702a86564a5bae9c0fee0d212d688 Mar 17 15:58:37 crc kubenswrapper[4767]: W0317 15:58:37.093496 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43b467d2_9860_4feb_a656_65827836c23e.slice/crio-298be9f25f706124d0ae69c87bc6e670590d8beb741c3edc02db29884c0fb298 WatchSource:0}: Error finding container 298be9f25f706124d0ae69c87bc6e670590d8beb741c3edc02db29884c0fb298: Status 404 returned error can't find the container with id 298be9f25f706124d0ae69c87bc6e670590d8beb741c3edc02db29884c0fb298 Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.197244 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-xfv84"] Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.257333 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96"] Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.279011 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c"] Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.312261 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-44q8c"] Mar 17 15:58:37 crc kubenswrapper[4767]: W0317 15:58:37.323961 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bfc70f3_ead2_43cc_9e90_7df32804d9ac.slice/crio-f0b34553e8534538ce09b40f0a25d967a638c65116fda210fb4210c3aecd969d WatchSource:0}: Error finding container f0b34553e8534538ce09b40f0a25d967a638c65116fda210fb4210c3aecd969d: Status 404 returned error can't find the container with id f0b34553e8534538ce09b40f0a25d967a638c65116fda210fb4210c3aecd969d Mar 17 15:58:37 crc kubenswrapper[4767]: W0317 15:58:37.331571 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda833dc16_3a29_4129_b592_732d71818bc4.slice/crio-ddc9b583fca160bab2e345d3bb8853c5fe6074916e904b3c7336124b24607df2 WatchSource:0}: Error finding container ddc9b583fca160bab2e345d3bb8853c5fe6074916e904b3c7336124b24607df2: Status 404 returned error can't find the container with id ddc9b583fca160bab2e345d3bb8853c5fe6074916e904b3c7336124b24607df2 Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.340043 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-btzsh"] Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.509263 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" event={"ID":"904d9f3b-95f0-4e57-8d04-c2026227a4a6","Type":"ContainerStarted","Data":"09d9a18f3008bb7ca7c85dbf742651de9913a46fff7915d91fd59a63031e5807"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.511293 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" event={"ID":"81201888-f7c3-4605-86a7-85f3edbca3a6","Type":"ContainerStarted","Data":"711a9fef8d465eade519fce16df7115793e5b72bc6b333757959f2358cd1641a"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.513043 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" event={"ID":"a833dc16-3a29-4129-b592-732d71818bc4","Type":"ContainerStarted","Data":"ddc9b583fca160bab2e345d3bb8853c5fe6074916e904b3c7336124b24607df2"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.515778 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" event={"ID":"fa1e1603-adc4-4ade-aca8-e59df8c350b2","Type":"ContainerStarted","Data":"d4db91a411dbc225ba09272f950b16174bef13bc30bda06d78194b0cd38b62bd"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.525050 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" event={"ID":"921c5e77-9858-4177-99a0-8cd3a4420d7d","Type":"ContainerStarted","Data":"700cecef7c75a42c83d31a2df4b9ac62d1e179bd8159ed55763626014eda820f"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.526881 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" event={"ID":"43b467d2-9860-4feb-a656-65827836c23e","Type":"ContainerStarted","Data":"298be9f25f706124d0ae69c87bc6e670590d8beb741c3edc02db29884c0fb298"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.530704 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" event={"ID":"a67a8ddc-0801-4530-bf82-9d4f4d0389a8","Type":"ContainerStarted","Data":"081b1095b3c6a82b60bb08e58ef4958d6585ec9d90dc392629bc9104ad068367"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.532668 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" event={"ID":"0163654c-d57e-4b14-aba0-f76dbaff1114","Type":"ContainerStarted","Data":"897825e540d75f3ba17f08a6033a0abcd873b492b0d899173f2980798a58ddbb"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.541869 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" event={"ID":"a7b5a1f5-da92-46dd-a5b0-5088e75346af","Type":"ContainerStarted","Data":"3ce3fd432023c5fa20121208cd9aa992c13e977137e9802ecf37b49e2ad893c8"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.546275 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" event={"ID":"16e4d9a1-285b-4221-8a99-55d515bc3356","Type":"ContainerStarted","Data":"fc6093cec954ce058e9c9e44b0afec26df9eb5c7c46221b81bbe7993802372fe"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.549642 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" event={"ID":"8bfc70f3-ead2-43cc-9e90-7df32804d9ac","Type":"ContainerStarted","Data":"f0b34553e8534538ce09b40f0a25d967a638c65116fda210fb4210c3aecd969d"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.556600 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" event={"ID":"bd2dbd72-69bf-40c1-b591-be3782c33465","Type":"ContainerStarted","Data":"ce2737a2768cc0c9e765ade2b2c212e038ddb30c0e8cf148ce9cab3bf2290067"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.565972 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" event={"ID":"59ba3b25-1e95-41ad-921c-9ee4ec5e2c43","Type":"ContainerStarted","Data":"1a748f50ca8fcb8f1ac7b4cdc59d946f07b702a86564a5bae9c0fee0d212d688"} Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.606160 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q"] Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.626239 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r"] Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.643583 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8"] Mar 17 15:58:37 crc kubenswrapper[4767]: W0317 15:58:37.644694 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7381ff68_5e5d_4281_9924_8495ece760f4.slice/crio-9e014cc3cd87d7965dfce8d10366e6c5c0d622bda6d934e0dfbd717257a1a035 WatchSource:0}: Error finding container 9e014cc3cd87d7965dfce8d10366e6c5c0d622bda6d934e0dfbd717257a1a035: Status 404 returned error can't find the container with id 9e014cc3cd87d7965dfce8d10366e6c5c0d622bda6d934e0dfbd717257a1a035 Mar 17 15:58:37 crc kubenswrapper[4767]: W0317 15:58:37.646287 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfaf77dfd_c76e_4abd_a8c8_b6211cf070a4.slice/crio-c39feca43f05552ebe7dcadc2449041e26487feda8641677937c939c751ee48f WatchSource:0}: Error finding container c39feca43f05552ebe7dcadc2449041e26487feda8641677937c939c751ee48f: Status 404 returned error can't find the container with id c39feca43f05552ebe7dcadc2449041e26487feda8641677937c939c751ee48f Mar 17 15:58:37 crc kubenswrapper[4767]: W0317 15:58:37.649290 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod248e2576_adbb_40bb_ade4_8c426f2a8a48.slice/crio-783dab42d4aa5c04d37d92a7edac0ce05dd890140d5604ae89113035defc86a1 WatchSource:0}: Error finding container 783dab42d4aa5c04d37d92a7edac0ce05dd890140d5604ae89113035defc86a1: Status 404 returned error can't find the container with id 783dab42d4aa5c04d37d92a7edac0ce05dd890140d5604ae89113035defc86a1 Mar 17 15:58:37 crc kubenswrapper[4767]: E0317 15:58:37.652935 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jl9fj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-tvpkb_openstack-operators(248e2576-adbb-40bb-ade4-8c426f2a8a48): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.653517 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb"] Mar 17 15:58:37 crc kubenswrapper[4767]: E0317 15:58:37.654320 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb" podUID="248e2576-adbb-40bb-ade4-8c426f2a8a48" Mar 17 15:58:37 crc kubenswrapper[4767]: W0317 15:58:37.661423 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad853953_ec48_40fc_8787_b2b838c955e9.slice/crio-7be3e06c5e77b1c8dbe82eb181048aec1d13316d50e2467cf90faecbf53c21ea WatchSource:0}: Error finding container 7be3e06c5e77b1c8dbe82eb181048aec1d13316d50e2467cf90faecbf53c21ea: Status 404 returned error can't find the container with id 7be3e06c5e77b1c8dbe82eb181048aec1d13316d50e2467cf90faecbf53c21ea Mar 17 15:58:37 crc kubenswrapper[4767]: I0317 15:58:37.664138 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5"] Mar 17 15:58:37 crc kubenswrapper[4767]: E0317 15:58:37.671993 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b4wp7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-c674c5965-jx2k5_openstack-operators(ad853953-ec48-40fc-8787-b2b838c955e9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 17 15:58:37 crc kubenswrapper[4767]: E0317 15:58:37.673144 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" podUID="ad853953-ec48-40fc-8787-b2b838c955e9" Mar 17 15:58:37 crc kubenswrapper[4767]: E0317 15:58:37.675828 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b4psx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-b8f8r_openstack-operators(e59ff072-51c2-4995-9f61-709f12a1393a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 17 15:58:37 crc kubenswrapper[4767]: E0317 15:58:37.677025 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" podUID="e59ff072-51c2-4995-9f61-709f12a1393a" Mar 17 15:58:38 crc kubenswrapper[4767]: I0317 15:58:38.269692 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-pjldm\" (UID: \"70fbe8aa-8647-4fe8-914c-0c05399cf46d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:38 crc kubenswrapper[4767]: E0317 15:58:38.270489 4767 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 15:58:38 crc kubenswrapper[4767]: E0317 15:58:38.270566 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert podName:70fbe8aa-8647-4fe8-914c-0c05399cf46d nodeName:}" failed. No retries permitted until 2026-03-17 15:58:42.270542893 +0000 UTC m=+1313.683858940 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-pjldm" (UID: "70fbe8aa-8647-4fe8-914c-0c05399cf46d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 15:58:38 crc kubenswrapper[4767]: I0317 15:58:38.476418 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:38 crc kubenswrapper[4767]: I0317 15:58:38.476894 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:38 crc kubenswrapper[4767]: E0317 15:58:38.477416 4767 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 17 15:58:38 crc kubenswrapper[4767]: E0317 15:58:38.477492 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs podName:713154c5-7e16-498f-9612-1c0afbf362bb nodeName:}" failed. No retries permitted until 2026-03-17 15:58:42.477470947 +0000 UTC m=+1313.890786994 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs") pod "openstack-operator-controller-manager-d7dccc75b-zlss6" (UID: "713154c5-7e16-498f-9612-1c0afbf362bb") : secret "webhook-server-cert" not found Mar 17 15:58:38 crc kubenswrapper[4767]: E0317 15:58:38.477553 4767 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 17 15:58:38 crc kubenswrapper[4767]: E0317 15:58:38.477577 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs podName:713154c5-7e16-498f-9612-1c0afbf362bb nodeName:}" failed. No retries permitted until 2026-03-17 15:58:42.477568779 +0000 UTC m=+1313.890884826 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs") pod "openstack-operator-controller-manager-d7dccc75b-zlss6" (UID: "713154c5-7e16-498f-9612-1c0afbf362bb") : secret "metrics-server-cert" not found Mar 17 15:58:38 crc kubenswrapper[4767]: I0317 15:58:38.609960 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" event={"ID":"ad853953-ec48-40fc-8787-b2b838c955e9","Type":"ContainerStarted","Data":"7be3e06c5e77b1c8dbe82eb181048aec1d13316d50e2467cf90faecbf53c21ea"} Mar 17 15:58:38 crc kubenswrapper[4767]: E0317 15:58:38.614257 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" podUID="ad853953-ec48-40fc-8787-b2b838c955e9" Mar 17 15:58:38 crc kubenswrapper[4767]: I0317 15:58:38.614874 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" event={"ID":"7381ff68-5e5d-4281-9924-8495ece760f4","Type":"ContainerStarted","Data":"9e014cc3cd87d7965dfce8d10366e6c5c0d622bda6d934e0dfbd717257a1a035"} Mar 17 15:58:38 crc kubenswrapper[4767]: I0317 15:58:38.617354 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb" event={"ID":"248e2576-adbb-40bb-ade4-8c426f2a8a48","Type":"ContainerStarted","Data":"783dab42d4aa5c04d37d92a7edac0ce05dd890140d5604ae89113035defc86a1"} Mar 17 15:58:38 crc kubenswrapper[4767]: E0317 15:58:38.618969 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb" podUID="248e2576-adbb-40bb-ade4-8c426f2a8a48" Mar 17 15:58:38 crc kubenswrapper[4767]: I0317 15:58:38.620142 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" event={"ID":"faf77dfd-c76e-4abd-a8c8-b6211cf070a4","Type":"ContainerStarted","Data":"c39feca43f05552ebe7dcadc2449041e26487feda8641677937c939c751ee48f"} Mar 17 15:58:38 crc kubenswrapper[4767]: I0317 15:58:38.626320 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" event={"ID":"e59ff072-51c2-4995-9f61-709f12a1393a","Type":"ContainerStarted","Data":"ba173a18878d6916fbc5558fd63b9310c16a758958099ab6bcddf8f8fda81c35"} Mar 17 15:58:38 crc kubenswrapper[4767]: E0317 15:58:38.628985 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" podUID="e59ff072-51c2-4995-9f61-709f12a1393a" Mar 17 15:58:39 crc kubenswrapper[4767]: E0317 15:58:39.653481 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" podUID="e59ff072-51c2-4995-9f61-709f12a1393a" Mar 17 15:58:39 crc kubenswrapper[4767]: E0317 15:58:39.654145 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb" podUID="248e2576-adbb-40bb-ade4-8c426f2a8a48" Mar 17 15:58:39 crc kubenswrapper[4767]: E0317 15:58:39.654218 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" podUID="ad853953-ec48-40fc-8787-b2b838c955e9" Mar 17 15:58:40 crc kubenswrapper[4767]: E0317 15:58:40.726155 4767 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 17 15:58:40 crc kubenswrapper[4767]: E0317 15:58:40.726330 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert podName:95d5d3e8-dc72-414f-afe4-b68b757a39d4 nodeName:}" failed. No retries permitted until 2026-03-17 15:58:48.726304697 +0000 UTC m=+1320.139620744 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert") pod "infra-operator-controller-manager-7b9c774f96-kftt2" (UID: "95d5d3e8-dc72-414f-afe4-b68b757a39d4") : secret "infra-operator-webhook-server-cert" not found Mar 17 15:58:40 crc kubenswrapper[4767]: I0317 15:58:40.725643 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert\") pod \"infra-operator-controller-manager-7b9c774f96-kftt2\" (UID: \"95d5d3e8-dc72-414f-afe4-b68b757a39d4\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:42 crc kubenswrapper[4767]: I0317 15:58:42.285614 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-pjldm\" (UID: \"70fbe8aa-8647-4fe8-914c-0c05399cf46d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:42 crc kubenswrapper[4767]: E0317 15:58:42.286329 4767 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 15:58:42 crc kubenswrapper[4767]: E0317 15:58:42.286896 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert podName:70fbe8aa-8647-4fe8-914c-0c05399cf46d nodeName:}" failed. No retries permitted until 2026-03-17 15:58:50.286846187 +0000 UTC m=+1321.700162244 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-pjldm" (UID: "70fbe8aa-8647-4fe8-914c-0c05399cf46d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 15:58:42 crc kubenswrapper[4767]: I0317 15:58:42.786580 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:42 crc kubenswrapper[4767]: I0317 15:58:42.787022 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:42 crc kubenswrapper[4767]: E0317 15:58:42.787719 4767 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 17 15:58:42 crc kubenswrapper[4767]: E0317 15:58:42.787839 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs podName:713154c5-7e16-498f-9612-1c0afbf362bb nodeName:}" failed. No retries permitted until 2026-03-17 15:58:50.787803742 +0000 UTC m=+1322.201119789 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs") pod "openstack-operator-controller-manager-d7dccc75b-zlss6" (UID: "713154c5-7e16-498f-9612-1c0afbf362bb") : secret "metrics-server-cert" not found Mar 17 15:58:42 crc kubenswrapper[4767]: E0317 15:58:42.788605 4767 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 17 15:58:42 crc kubenswrapper[4767]: E0317 15:58:42.788664 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs podName:713154c5-7e16-498f-9612-1c0afbf362bb nodeName:}" failed. No retries permitted until 2026-03-17 15:58:50.788649904 +0000 UTC m=+1322.201965951 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs") pod "openstack-operator-controller-manager-d7dccc75b-zlss6" (UID: "713154c5-7e16-498f-9612-1c0afbf362bb") : secret "webhook-server-cert" not found Mar 17 15:58:48 crc kubenswrapper[4767]: I0317 15:58:48.743732 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert\") pod \"infra-operator-controller-manager-7b9c774f96-kftt2\" (UID: \"95d5d3e8-dc72-414f-afe4-b68b757a39d4\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:48 crc kubenswrapper[4767]: I0317 15:58:48.753499 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/95d5d3e8-dc72-414f-afe4-b68b757a39d4-cert\") pod \"infra-operator-controller-manager-7b9c774f96-kftt2\" (UID: \"95d5d3e8-dc72-414f-afe4-b68b757a39d4\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:48 crc kubenswrapper[4767]: I0317 15:58:48.865366 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:58:49 crc kubenswrapper[4767]: E0317 15:58:49.554656 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad" Mar 17 15:58:49 crc kubenswrapper[4767]: E0317 15:58:49.555355 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjgx4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-588d4d986b-q7zd5_openstack-operators(a67a8ddc-0801-4530-bf82-9d4f4d0389a8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 15:58:49 crc kubenswrapper[4767]: E0317 15:58:49.556588 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" podUID="a67a8ddc-0801-4530-bf82-9d4f4d0389a8" Mar 17 15:58:49 crc kubenswrapper[4767]: E0317 15:58:49.883438 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad\\\"\"" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" podUID="a67a8ddc-0801-4530-bf82-9d4f4d0389a8" Mar 17 15:58:50 crc kubenswrapper[4767]: I0317 15:58:50.340429 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-pjldm\" (UID: \"70fbe8aa-8647-4fe8-914c-0c05399cf46d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:50 crc kubenswrapper[4767]: I0317 15:58:50.396146 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70fbe8aa-8647-4fe8-914c-0c05399cf46d-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-pjldm\" (UID: \"70fbe8aa-8647-4fe8-914c-0c05399cf46d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:50 crc kubenswrapper[4767]: I0317 15:58:50.445612 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-hwtx7" Mar 17 15:58:50 crc kubenswrapper[4767]: I0317 15:58:50.451867 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:58:50 crc kubenswrapper[4767]: I0317 15:58:50.852701 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:50 crc kubenswrapper[4767]: I0317 15:58:50.852902 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:50 crc kubenswrapper[4767]: I0317 15:58:50.857795 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-webhook-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:50 crc kubenswrapper[4767]: I0317 15:58:50.863919 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/713154c5-7e16-498f-9612-1c0afbf362bb-metrics-certs\") pod \"openstack-operator-controller-manager-d7dccc75b-zlss6\" (UID: \"713154c5-7e16-498f-9612-1c0afbf362bb\") " pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:51 crc kubenswrapper[4767]: I0317 15:58:51.060745 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-f4m4f" Mar 17 15:58:51 crc kubenswrapper[4767]: I0317 15:58:51.068308 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:58:51 crc kubenswrapper[4767]: E0317 15:58:51.099921 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113" Mar 17 15:58:51 crc kubenswrapper[4767]: E0317 15:58:51.100209 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l84l2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-8464cc45fb-wj49c_openstack-operators(fa1e1603-adc4-4ade-aca8-e59df8c350b2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 15:58:51 crc kubenswrapper[4767]: E0317 15:58:51.101681 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" podUID="fa1e1603-adc4-4ade-aca8-e59df8c350b2" Mar 17 15:58:51 crc kubenswrapper[4767]: E0317 15:58:51.878230 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:7562d3e09bdac17f447f4523c5bd784c5f5ab5ca9cb2370a03b86126d6d7301d" Mar 17 15:58:51 crc kubenswrapper[4767]: E0317 15:58:51.878571 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:7562d3e09bdac17f447f4523c5bd784c5f5ab5ca9cb2370a03b86126d6d7301d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mmq67,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-59bc569d95-rvkxc_openstack-operators(53aea289-5a28-438b-8d28-242d836351f0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 15:58:51 crc kubenswrapper[4767]: E0317 15:58:51.879845 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" podUID="53aea289-5a28-438b-8d28-242d836351f0" Mar 17 15:58:51 crc kubenswrapper[4767]: E0317 15:58:51.907237 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" podUID="fa1e1603-adc4-4ade-aca8-e59df8c350b2" Mar 17 15:58:51 crc kubenswrapper[4767]: E0317 15:58:51.907325 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:7562d3e09bdac17f447f4523c5bd784c5f5ab5ca9cb2370a03b86126d6d7301d\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" podUID="53aea289-5a28-438b-8d28-242d836351f0" Mar 17 15:58:52 crc kubenswrapper[4767]: E0317 15:58:52.528588 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8" Mar 17 15:58:52 crc kubenswrapper[4767]: E0317 15:58:52.529988 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fc8m5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6f787dddc9-24wkj_openstack-operators(904d9f3b-95f0-4e57-8d04-c2026227a4a6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 15:58:52 crc kubenswrapper[4767]: E0317 15:58:52.531304 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" podUID="904d9f3b-95f0-4e57-8d04-c2026227a4a6" Mar 17 15:58:52 crc kubenswrapper[4767]: E0317 15:58:52.919328 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" podUID="904d9f3b-95f0-4e57-8d04-c2026227a4a6" Mar 17 15:58:53 crc kubenswrapper[4767]: E0317 15:58:53.957308 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d" Mar 17 15:58:53 crc kubenswrapper[4767]: E0317 15:58:53.958100 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5xcrq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-79df6bcc97-p2skw_openstack-operators(16e4d9a1-285b-4221-8a99-55d515bc3356): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 15:58:53 crc kubenswrapper[4767]: E0317 15:58:53.959364 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" podUID="16e4d9a1-285b-4221-8a99-55d515bc3356" Mar 17 15:58:54 crc kubenswrapper[4767]: E0317 15:58:54.875290 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a" Mar 17 15:58:54 crc kubenswrapper[4767]: E0317 15:58:54.875548 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j4xfl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5b9f45d989-xms5z_openstack-operators(59ba3b25-1e95-41ad-921c-9ee4ec5e2c43): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 15:58:54 crc kubenswrapper[4767]: E0317 15:58:54.876989 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" podUID="59ba3b25-1e95-41ad-921c-9ee4ec5e2c43" Mar 17 15:58:54 crc kubenswrapper[4767]: E0317 15:58:54.936718 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" podUID="59ba3b25-1e95-41ad-921c-9ee4ec5e2c43" Mar 17 15:58:54 crc kubenswrapper[4767]: E0317 15:58:54.936749 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d\\\"\"" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" podUID="16e4d9a1-285b-4221-8a99-55d515bc3356" Mar 17 15:58:56 crc kubenswrapper[4767]: E0317 15:58:56.081628 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622" Mar 17 15:58:56 crc kubenswrapper[4767]: E0317 15:58:56.082437 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zlc7d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5784578c99-2k2dh_openstack-operators(bd2dbd72-69bf-40c1-b591-be3782c33465): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 15:58:56 crc kubenswrapper[4767]: E0317 15:58:56.083685 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" podUID="bd2dbd72-69bf-40c1-b591-be3782c33465" Mar 17 15:58:56 crc kubenswrapper[4767]: E0317 15:58:56.752375 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a" Mar 17 15:58:56 crc kubenswrapper[4767]: E0317 15:58:56.753020 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqqn6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5d488d59fb-r8z96_openstack-operators(921c5e77-9858-4177-99a0-8cd3a4420d7d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 15:58:56 crc kubenswrapper[4767]: E0317 15:58:56.754336 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" podUID="921c5e77-9858-4177-99a0-8cd3a4420d7d" Mar 17 15:58:56 crc kubenswrapper[4767]: E0317 15:58:56.953725 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" podUID="bd2dbd72-69bf-40c1-b591-be3782c33465" Mar 17 15:58:56 crc kubenswrapper[4767]: E0317 15:58:56.955198 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" podUID="921c5e77-9858-4177-99a0-8cd3a4420d7d" Mar 17 15:58:58 crc kubenswrapper[4767]: E0317 15:58:58.579634 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.58:5001/openstack-k8s-operators/telemetry-operator:15c2ffcfe08e13a1dec28232b4ee653042564ac3" Mar 17 15:58:58 crc kubenswrapper[4767]: E0317 15:58:58.580162 4767 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.58:5001/openstack-k8s-operators/telemetry-operator:15c2ffcfe08e13a1dec28232b4ee653042564ac3" Mar 17 15:58:58 crc kubenswrapper[4767]: E0317 15:58:58.580398 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.58:5001/openstack-k8s-operators/telemetry-operator:15c2ffcfe08e13a1dec28232b4ee653042564ac3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mvf7z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-766dd46678-n9j9q_openstack-operators(7381ff68-5e5d-4281-9924-8495ece760f4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 15:58:58 crc kubenswrapper[4767]: E0317 15:58:58.581930 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" podUID="7381ff68-5e5d-4281-9924-8495ece760f4" Mar 17 15:58:58 crc kubenswrapper[4767]: E0317 15:58:58.976615 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.58:5001/openstack-k8s-operators/telemetry-operator:15c2ffcfe08e13a1dec28232b4ee653042564ac3\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" podUID="7381ff68-5e5d-4281-9924-8495ece760f4" Mar 17 15:58:59 crc kubenswrapper[4767]: E0317 15:58:59.179571 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da" Mar 17 15:58:59 crc kubenswrapper[4767]: E0317 15:58:59.179887 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zh9d5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-55f864c847-xfv84_openstack-operators(81201888-f7c3-4605-86a7-85f3edbca3a6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 15:58:59 crc kubenswrapper[4767]: E0317 15:58:59.181362 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" podUID="81201888-f7c3-4605-86a7-85f3edbca3a6" Mar 17 15:58:59 crc kubenswrapper[4767]: E0317 15:58:59.728038 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55" Mar 17 15:58:59 crc kubenswrapper[4767]: E0317 15:58:59.728281 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h2dsb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-884679f54-44q8c_openstack-operators(a833dc16-3a29-4129-b592-732d71818bc4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 15:58:59 crc kubenswrapper[4767]: E0317 15:58:59.729368 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" podUID="a833dc16-3a29-4129-b592-732d71818bc4" Mar 17 15:58:59 crc kubenswrapper[4767]: E0317 15:58:59.988390 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" podUID="a833dc16-3a29-4129-b592-732d71818bc4" Mar 17 15:58:59 crc kubenswrapper[4767]: E0317 15:58:59.989879 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da\\\"\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" podUID="81201888-f7c3-4605-86a7-85f3edbca3a6" Mar 17 15:59:04 crc kubenswrapper[4767]: W0317 15:59:04.922975 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70fbe8aa_8647_4fe8_914c_0c05399cf46d.slice/crio-fc4d6877b4481740d603cef2ee7a743b44e76bb8eb2240cb59e7f30c8202759d WatchSource:0}: Error finding container fc4d6877b4481740d603cef2ee7a743b44e76bb8eb2240cb59e7f30c8202759d: Status 404 returned error can't find the container with id fc4d6877b4481740d603cef2ee7a743b44e76bb8eb2240cb59e7f30c8202759d Mar 17 15:59:04 crc kubenswrapper[4767]: I0317 15:59:04.924241 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm"] Mar 17 15:59:05 crc kubenswrapper[4767]: I0317 15:59:05.015443 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2"] Mar 17 15:59:05 crc kubenswrapper[4767]: I0317 15:59:05.024468 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6"] Mar 17 15:59:05 crc kubenswrapper[4767]: W0317 15:59:05.025070 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod713154c5_7e16_498f_9612_1c0afbf362bb.slice/crio-ac627947d9afee1ef566ecb44d3a3c8474fddb33f9a930f2ec972d5a16d99df1 WatchSource:0}: Error finding container ac627947d9afee1ef566ecb44d3a3c8474fddb33f9a930f2ec972d5a16d99df1: Status 404 returned error can't find the container with id ac627947d9afee1ef566ecb44d3a3c8474fddb33f9a930f2ec972d5a16d99df1 Mar 17 15:59:05 crc kubenswrapper[4767]: I0317 15:59:05.042314 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" event={"ID":"713154c5-7e16-498f-9612-1c0afbf362bb","Type":"ContainerStarted","Data":"ac627947d9afee1ef566ecb44d3a3c8474fddb33f9a930f2ec972d5a16d99df1"} Mar 17 15:59:05 crc kubenswrapper[4767]: I0317 15:59:05.044116 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" event={"ID":"a7b5a1f5-da92-46dd-a5b0-5088e75346af","Type":"ContainerStarted","Data":"e0db98d18a0069ffcf1f782be7c5223dd0bbf104fd9ff358d8c6daac1bd5edef"} Mar 17 15:59:05 crc kubenswrapper[4767]: I0317 15:59:05.044412 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" Mar 17 15:59:05 crc kubenswrapper[4767]: I0317 15:59:05.046028 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" event={"ID":"70fbe8aa-8647-4fe8-914c-0c05399cf46d","Type":"ContainerStarted","Data":"fc4d6877b4481740d603cef2ee7a743b44e76bb8eb2240cb59e7f30c8202759d"} Mar 17 15:59:05 crc kubenswrapper[4767]: I0317 15:59:05.047871 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" event={"ID":"95d5d3e8-dc72-414f-afe4-b68b757a39d4","Type":"ContainerStarted","Data":"7030a8930f61a859e3134a0899467e435dd26cc528c017ec3faccf70380f5a92"} Mar 17 15:59:05 crc kubenswrapper[4767]: I0317 15:59:05.068211 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" podStartSLOduration=9.893941273 podStartE2EDuration="33.06818664s" podCreationTimestamp="2026-03-17 15:58:32 +0000 UTC" firstStartedPulling="2026-03-17 15:58:36.576059671 +0000 UTC m=+1307.989375718" lastFinishedPulling="2026-03-17 15:58:59.750305038 +0000 UTC m=+1331.163621085" observedRunningTime="2026-03-17 15:59:05.063818757 +0000 UTC m=+1336.477134814" watchObservedRunningTime="2026-03-17 15:59:05.06818664 +0000 UTC m=+1336.481502687" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.147736 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" event={"ID":"8bfc70f3-ead2-43cc-9e90-7df32804d9ac","Type":"ContainerStarted","Data":"8e38134de4b8aac5b4698affeee017eb99dd1861726441fca8a6988a5eb544f0"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.150025 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.173422 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" event={"ID":"e59ff072-51c2-4995-9f61-709f12a1393a","Type":"ContainerStarted","Data":"40659e8a7dfbbdabdb7c5244845ba76a395f7824954f40eef8741026a2be3fe5"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.178728 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.181687 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" event={"ID":"fa1e1603-adc4-4ade-aca8-e59df8c350b2","Type":"ContainerStarted","Data":"9d31322b656504fd5d6e8cbee9840dd28aaccf38841b1289bff76c7aeca5418c"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.182514 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.232264 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" event={"ID":"a67a8ddc-0801-4530-bf82-9d4f4d0389a8","Type":"ContainerStarted","Data":"db0395eb96ca4956ac21888ab5778fc523e2737fa71b3ff83476194dd76f66b6"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.232874 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.243948 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" podStartSLOduration=10.824856433 podStartE2EDuration="33.243914336s" podCreationTimestamp="2026-03-17 15:58:33 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.331761479 +0000 UTC m=+1308.745077526" lastFinishedPulling="2026-03-17 15:58:59.750819382 +0000 UTC m=+1331.164135429" observedRunningTime="2026-03-17 15:59:06.204119615 +0000 UTC m=+1337.617435682" watchObservedRunningTime="2026-03-17 15:59:06.243914336 +0000 UTC m=+1337.657230383" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.277612 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb" event={"ID":"248e2576-adbb-40bb-ade4-8c426f2a8a48","Type":"ContainerStarted","Data":"25fb5b028026e27b91a36a40de7f1bfa427a1f4b9066c3125afd2cfd3db244f1"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.316136 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" podStartSLOduration=6.992225409 podStartE2EDuration="34.316100307s" podCreationTimestamp="2026-03-17 15:58:32 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.276009264 +0000 UTC m=+1308.689325311" lastFinishedPulling="2026-03-17 15:59:04.599884162 +0000 UTC m=+1336.013200209" observedRunningTime="2026-03-17 15:59:06.262309993 +0000 UTC m=+1337.675626050" watchObservedRunningTime="2026-03-17 15:59:06.316100307 +0000 UTC m=+1337.729416354" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.316513 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" event={"ID":"b6dc3b02-8dc0-46ac-b78e-1d7d3ffb3f24","Type":"ContainerStarted","Data":"881489395a179c513730a8d0aea9bd69be200d5701c49649fb6984ea78506ff1"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.317729 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.332472 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" event={"ID":"53aea289-5a28-438b-8d28-242d836351f0","Type":"ContainerStarted","Data":"8170327490f5dcbc185f97b00164d96a2c592f55455089c7fca371b671bd7ee4"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.332819 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.352415 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" event={"ID":"faf77dfd-c76e-4abd-a8c8-b6211cf070a4","Type":"ContainerStarted","Data":"12befc50e89584186a8f6c22f761c726609b7d870ce1c511dcadaa5cac6a6ed4"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.353671 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.359736 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" podStartSLOduration=6.608000571 podStartE2EDuration="33.359700417s" podCreationTimestamp="2026-03-17 15:58:33 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.675101959 +0000 UTC m=+1309.088418006" lastFinishedPulling="2026-03-17 15:59:04.426801805 +0000 UTC m=+1335.840117852" observedRunningTime="2026-03-17 15:59:06.301734655 +0000 UTC m=+1337.715051152" watchObservedRunningTime="2026-03-17 15:59:06.359700417 +0000 UTC m=+1337.773016464" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.394192 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" event={"ID":"ad853953-ec48-40fc-8787-b2b838c955e9","Type":"ContainerStarted","Data":"a9e6ce7f2c3ca1a48e1fc7579ff6c43f30c15562c53c4d0759378ffe379e0c5e"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.394150 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tvpkb" podStartSLOduration=5.540051439 podStartE2EDuration="32.394109789s" podCreationTimestamp="2026-03-17 15:58:34 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.652746679 +0000 UTC m=+1309.066062726" lastFinishedPulling="2026-03-17 15:59:04.506805019 +0000 UTC m=+1335.920121076" observedRunningTime="2026-03-17 15:59:06.349046171 +0000 UTC m=+1337.762362218" watchObservedRunningTime="2026-03-17 15:59:06.394109789 +0000 UTC m=+1337.807425846" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.395273 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.406034 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" event={"ID":"43b467d2-9860-4feb-a656-65827836c23e","Type":"ContainerStarted","Data":"8b10143429a4ed2f968228b6412731cafb129ea94228f7e10753549de978fe75"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.407063 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.415290 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" podStartSLOduration=6.462358215 podStartE2EDuration="34.415158135s" podCreationTimestamp="2026-03-17 15:58:32 +0000 UTC" firstStartedPulling="2026-03-17 15:58:36.649762701 +0000 UTC m=+1308.063078748" lastFinishedPulling="2026-03-17 15:59:04.602562621 +0000 UTC m=+1336.015878668" observedRunningTime="2026-03-17 15:59:06.388514834 +0000 UTC m=+1337.801830891" watchObservedRunningTime="2026-03-17 15:59:06.415158135 +0000 UTC m=+1337.828474172" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.441678 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" event={"ID":"0163654c-d57e-4b14-aba0-f76dbaff1114","Type":"ContainerStarted","Data":"93043afec3364398b137dadb59356bc93ac1dc7c23776f368311e06c0ecf6c20"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.441878 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.443108 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" podStartSLOduration=9.242083168 podStartE2EDuration="34.443081089s" podCreationTimestamp="2026-03-17 15:58:32 +0000 UTC" firstStartedPulling="2026-03-17 15:58:34.521879186 +0000 UTC m=+1305.935195233" lastFinishedPulling="2026-03-17 15:58:59.722877107 +0000 UTC m=+1331.136193154" observedRunningTime="2026-03-17 15:59:06.429414724 +0000 UTC m=+1337.842730871" watchObservedRunningTime="2026-03-17 15:59:06.443081089 +0000 UTC m=+1337.856397156" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.446132 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" event={"ID":"713154c5-7e16-498f-9612-1c0afbf362bb","Type":"ContainerStarted","Data":"74b92aebeee36515abe632c0cc948e130b050afb7eea4c63dbaa839fecf70d3d"} Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.471653 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" podStartSLOduration=7.613528924 podStartE2EDuration="33.471627809s" podCreationTimestamp="2026-03-17 15:58:33 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.651481886 +0000 UTC m=+1309.064797933" lastFinishedPulling="2026-03-17 15:59:03.509580781 +0000 UTC m=+1334.922896818" observedRunningTime="2026-03-17 15:59:06.471419133 +0000 UTC m=+1337.884735190" watchObservedRunningTime="2026-03-17 15:59:06.471627809 +0000 UTC m=+1337.884943856" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.588355 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" podStartSLOduration=4.216397758 podStartE2EDuration="34.588319563s" podCreationTimestamp="2026-03-17 15:58:32 +0000 UTC" firstStartedPulling="2026-03-17 15:58:34.810928198 +0000 UTC m=+1306.224244235" lastFinishedPulling="2026-03-17 15:59:05.182849993 +0000 UTC m=+1336.596166040" observedRunningTime="2026-03-17 15:59:06.565692927 +0000 UTC m=+1337.979008984" watchObservedRunningTime="2026-03-17 15:59:06.588319563 +0000 UTC m=+1338.001635610" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.646150 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" podStartSLOduration=6.890681517 podStartE2EDuration="33.646102921s" podCreationTimestamp="2026-03-17 15:58:33 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.67169781 +0000 UTC m=+1309.085013847" lastFinishedPulling="2026-03-17 15:59:04.427119204 +0000 UTC m=+1335.840435251" observedRunningTime="2026-03-17 15:59:06.615552319 +0000 UTC m=+1338.028868386" watchObservedRunningTime="2026-03-17 15:59:06.646102921 +0000 UTC m=+1338.059418968" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.689532 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" podStartSLOduration=10.048915672 podStartE2EDuration="34.689491016s" podCreationTimestamp="2026-03-17 15:58:32 +0000 UTC" firstStartedPulling="2026-03-17 15:58:36.602752263 +0000 UTC m=+1308.016068310" lastFinishedPulling="2026-03-17 15:59:01.243327607 +0000 UTC m=+1332.656643654" observedRunningTime="2026-03-17 15:59:06.643275618 +0000 UTC m=+1338.056591685" watchObservedRunningTime="2026-03-17 15:59:06.689491016 +0000 UTC m=+1338.102807063" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.696137 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" podStartSLOduration=12.047246848 podStartE2EDuration="34.696107697s" podCreationTimestamp="2026-03-17 15:58:32 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.101286585 +0000 UTC m=+1308.514602632" lastFinishedPulling="2026-03-17 15:58:59.750147434 +0000 UTC m=+1331.163463481" observedRunningTime="2026-03-17 15:59:06.682783792 +0000 UTC m=+1338.096099849" watchObservedRunningTime="2026-03-17 15:59:06.696107697 +0000 UTC m=+1338.109423744" Mar 17 15:59:06 crc kubenswrapper[4767]: I0317 15:59:06.781338 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" podStartSLOduration=33.781286305 podStartE2EDuration="33.781286305s" podCreationTimestamp="2026-03-17 15:58:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 15:59:06.754220744 +0000 UTC m=+1338.167536801" watchObservedRunningTime="2026-03-17 15:59:06.781286305 +0000 UTC m=+1338.194602352" Mar 17 15:59:07 crc kubenswrapper[4767]: I0317 15:59:07.495281 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" event={"ID":"59ba3b25-1e95-41ad-921c-9ee4ec5e2c43","Type":"ContainerStarted","Data":"4d4b716faf520d9a4ca657e0d7d163e03e65ffeb116ebaedfbe80d11d0bfe322"} Mar 17 15:59:07 crc kubenswrapper[4767]: I0317 15:59:07.499231 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:59:07 crc kubenswrapper[4767]: I0317 15:59:07.499289 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" Mar 17 15:59:07 crc kubenswrapper[4767]: I0317 15:59:07.517746 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" podStartSLOduration=4.576441472 podStartE2EDuration="34.517713844s" podCreationTimestamp="2026-03-17 15:58:33 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.092946409 +0000 UTC m=+1308.506262456" lastFinishedPulling="2026-03-17 15:59:07.034218781 +0000 UTC m=+1338.447534828" observedRunningTime="2026-03-17 15:59:07.514569462 +0000 UTC m=+1338.927885519" watchObservedRunningTime="2026-03-17 15:59:07.517713844 +0000 UTC m=+1338.931029901" Mar 17 15:59:08 crc kubenswrapper[4767]: I0317 15:59:08.507266 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" event={"ID":"921c5e77-9858-4177-99a0-8cd3a4420d7d","Type":"ContainerStarted","Data":"204d75195a2f996056746fc3b56900ac3f45b8b912281c14865ffa9b0b1e339f"} Mar 17 15:59:08 crc kubenswrapper[4767]: I0317 15:59:08.532357 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" podStartSLOduration=4.855239428 podStartE2EDuration="35.532331203s" podCreationTimestamp="2026-03-17 15:58:33 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.328255118 +0000 UTC m=+1308.741571165" lastFinishedPulling="2026-03-17 15:59:08.005346893 +0000 UTC m=+1339.418662940" observedRunningTime="2026-03-17 15:59:08.527761944 +0000 UTC m=+1339.941078011" watchObservedRunningTime="2026-03-17 15:59:08.532331203 +0000 UTC m=+1339.945647240" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.080131 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.775405 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" event={"ID":"bd2dbd72-69bf-40c1-b591-be3782c33465","Type":"ContainerStarted","Data":"98e1bcbc139696e833afa6fe1b4306d8435137f8d3f8c965a1d0672b2f28d30c"} Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.776116 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.795819 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" event={"ID":"81201888-f7c3-4605-86a7-85f3edbca3a6","Type":"ContainerStarted","Data":"3b563dfbd235a4ab47726013468dc87e8cf7223f2e89044a104826f7530dc17e"} Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.797058 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.809611 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" event={"ID":"70fbe8aa-8647-4fe8-914c-0c05399cf46d","Type":"ContainerStarted","Data":"f42bd3c4f2af2dd50de480c34d99956b8b583c9a3551b160871025ce4d7b5666"} Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.811456 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.818063 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" event={"ID":"95d5d3e8-dc72-414f-afe4-b68b757a39d4","Type":"ContainerStarted","Data":"96b93e28c5e8a81213a9c9526312a1cc8018d14134c4769587830880604b5db3"} Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.818479 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.821814 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" event={"ID":"904d9f3b-95f0-4e57-8d04-c2026227a4a6","Type":"ContainerStarted","Data":"1e47fd04dd013933d9e3c724fbf14b32d6b4c2a21b0d2c1b23dd76b6abb46f2f"} Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.823278 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.824154 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" event={"ID":"16e4d9a1-285b-4221-8a99-55d515bc3356","Type":"ContainerStarted","Data":"6b29dbb58860b9bd664e49130231961d40d2ddeadf1ca17d48ddcdf542e3d7c7"} Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.824457 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.831919 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" podStartSLOduration=5.350016552 podStartE2EDuration="38.831887339s" podCreationTimestamp="2026-03-17 15:58:33 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.080117746 +0000 UTC m=+1308.493433783" lastFinishedPulling="2026-03-17 15:59:10.561988523 +0000 UTC m=+1341.975304570" observedRunningTime="2026-03-17 15:59:11.81111554 +0000 UTC m=+1343.224431617" watchObservedRunningTime="2026-03-17 15:59:11.831887339 +0000 UTC m=+1343.245203386" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.868017 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" podStartSLOduration=33.236188824 podStartE2EDuration="38.867978964s" podCreationTimestamp="2026-03-17 15:58:33 +0000 UTC" firstStartedPulling="2026-03-17 15:59:04.928493229 +0000 UTC m=+1336.341809276" lastFinishedPulling="2026-03-17 15:59:10.560283369 +0000 UTC m=+1341.973599416" observedRunningTime="2026-03-17 15:59:11.865147681 +0000 UTC m=+1343.278463758" watchObservedRunningTime="2026-03-17 15:59:11.867978964 +0000 UTC m=+1343.281295011" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.932119 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" podStartSLOduration=6.414153805 podStartE2EDuration="39.932094026s" podCreationTimestamp="2026-03-17 15:58:32 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.257909465 +0000 UTC m=+1308.671225512" lastFinishedPulling="2026-03-17 15:59:10.775849686 +0000 UTC m=+1342.189165733" observedRunningTime="2026-03-17 15:59:11.905908488 +0000 UTC m=+1343.319224565" watchObservedRunningTime="2026-03-17 15:59:11.932094026 +0000 UTC m=+1343.345410073" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.940708 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" podStartSLOduration=5.970846104 podStartE2EDuration="39.940683609s" podCreationTimestamp="2026-03-17 15:58:32 +0000 UTC" firstStartedPulling="2026-03-17 15:58:36.5937751 +0000 UTC m=+1308.007091147" lastFinishedPulling="2026-03-17 15:59:10.563612605 +0000 UTC m=+1341.976928652" observedRunningTime="2026-03-17 15:59:11.931423199 +0000 UTC m=+1343.344739246" watchObservedRunningTime="2026-03-17 15:59:11.940683609 +0000 UTC m=+1343.353999656" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.955390 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" podStartSLOduration=6.031413824 podStartE2EDuration="39.95536635s" podCreationTimestamp="2026-03-17 15:58:32 +0000 UTC" firstStartedPulling="2026-03-17 15:58:36.639373732 +0000 UTC m=+1308.052689789" lastFinishedPulling="2026-03-17 15:59:10.563326278 +0000 UTC m=+1341.976642315" observedRunningTime="2026-03-17 15:59:11.951374866 +0000 UTC m=+1343.364690923" watchObservedRunningTime="2026-03-17 15:59:11.95536635 +0000 UTC m=+1343.368682397" Mar 17 15:59:11 crc kubenswrapper[4767]: I0317 15:59:11.974808 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" podStartSLOduration=34.427217917 podStartE2EDuration="39.974773673s" podCreationTimestamp="2026-03-17 15:58:32 +0000 UTC" firstStartedPulling="2026-03-17 15:59:05.015140845 +0000 UTC m=+1336.428456892" lastFinishedPulling="2026-03-17 15:59:10.562696601 +0000 UTC m=+1341.976012648" observedRunningTime="2026-03-17 15:59:11.971148099 +0000 UTC m=+1343.384464146" watchObservedRunningTime="2026-03-17 15:59:11.974773673 +0000 UTC m=+1343.388089720" Mar 17 15:59:12 crc kubenswrapper[4767]: I0317 15:59:12.843353 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" event={"ID":"7381ff68-5e5d-4281-9924-8495ece760f4","Type":"ContainerStarted","Data":"57936bb04a1a7821e849b962a514d01de9996e9eba89adf6f2af8fe586da845b"} Mar 17 15:59:12 crc kubenswrapper[4767]: I0317 15:59:12.844623 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" Mar 17 15:59:12 crc kubenswrapper[4767]: I0317 15:59:12.910367 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" podStartSLOduration=5.121433208 podStartE2EDuration="39.910321523s" podCreationTimestamp="2026-03-17 15:58:33 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.650427499 +0000 UTC m=+1309.063743546" lastFinishedPulling="2026-03-17 15:59:12.439315814 +0000 UTC m=+1343.852631861" observedRunningTime="2026-03-17 15:59:12.889682098 +0000 UTC m=+1344.302998155" watchObservedRunningTime="2026-03-17 15:59:12.910321523 +0000 UTC m=+1344.323637570" Mar 17 15:59:12 crc kubenswrapper[4767]: I0317 15:59:12.964718 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" Mar 17 15:59:12 crc kubenswrapper[4767]: I0317 15:59:12.984107 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" Mar 17 15:59:13 crc kubenswrapper[4767]: I0317 15:59:13.310503 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" Mar 17 15:59:13 crc kubenswrapper[4767]: I0317 15:59:13.484016 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" Mar 17 15:59:13 crc kubenswrapper[4767]: I0317 15:59:13.512041 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" Mar 17 15:59:13 crc kubenswrapper[4767]: I0317 15:59:13.856664 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" event={"ID":"a833dc16-3a29-4129-b592-732d71818bc4","Type":"ContainerStarted","Data":"df8ef7becc2c6c9d2bcb595856a83a590570e0f9c8858bd10297702f2d1472b2"} Mar 17 15:59:13 crc kubenswrapper[4767]: I0317 15:59:13.858621 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" Mar 17 15:59:13 crc kubenswrapper[4767]: I0317 15:59:13.903915 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" podStartSLOduration=5.334122201 podStartE2EDuration="40.903871406s" podCreationTimestamp="2026-03-17 15:58:33 +0000 UTC" firstStartedPulling="2026-03-17 15:58:37.33836865 +0000 UTC m=+1308.751684697" lastFinishedPulling="2026-03-17 15:59:12.908117855 +0000 UTC m=+1344.321433902" observedRunningTime="2026-03-17 15:59:13.893730473 +0000 UTC m=+1345.307046540" watchObservedRunningTime="2026-03-17 15:59:13.903871406 +0000 UTC m=+1345.317187453" Mar 17 15:59:14 crc kubenswrapper[4767]: I0317 15:59:14.502909 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" Mar 17 15:59:14 crc kubenswrapper[4767]: I0317 15:59:14.515840 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" Mar 17 15:59:14 crc kubenswrapper[4767]: I0317 15:59:14.522357 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" Mar 17 15:59:14 crc kubenswrapper[4767]: I0317 15:59:14.553342 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" Mar 17 15:59:14 crc kubenswrapper[4767]: I0317 15:59:14.587234 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" Mar 17 15:59:14 crc kubenswrapper[4767]: I0317 15:59:14.812220 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" Mar 17 15:59:14 crc kubenswrapper[4767]: I0317 15:59:14.859947 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" Mar 17 15:59:14 crc kubenswrapper[4767]: I0317 15:59:14.868677 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" Mar 17 15:59:15 crc kubenswrapper[4767]: I0317 15:59:15.443776 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" Mar 17 15:59:18 crc kubenswrapper[4767]: I0317 15:59:18.876588 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 15:59:20 crc kubenswrapper[4767]: I0317 15:59:20.460624 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 15:59:23 crc kubenswrapper[4767]: I0317 15:59:23.489596 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" Mar 17 15:59:23 crc kubenswrapper[4767]: I0317 15:59:23.587838 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" Mar 17 15:59:24 crc kubenswrapper[4767]: I0317 15:59:24.451923 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" Mar 17 15:59:24 crc kubenswrapper[4767]: I0317 15:59:24.635747 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" Mar 17 15:59:24 crc kubenswrapper[4767]: I0317 15:59:24.685132 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" Mar 17 15:59:24 crc kubenswrapper[4767]: I0317 15:59:24.867104 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" Mar 17 15:59:34 crc kubenswrapper[4767]: I0317 15:59:34.166896 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 15:59:34 crc kubenswrapper[4767]: I0317 15:59:34.167824 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.019219 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7v6b4"] Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.022684 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.032620 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.032917 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-f4ttn" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.032952 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.042899 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.058957 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsvdr\" (UniqueName: \"kubernetes.io/projected/999a6b3f-d0b8-477e-8be5-73488425a861-kube-api-access-vsvdr\") pod \"dnsmasq-dns-675f4bcbfc-7v6b4\" (UID: \"999a6b3f-d0b8-477e-8be5-73488425a861\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.059320 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/999a6b3f-d0b8-477e-8be5-73488425a861-config\") pod \"dnsmasq-dns-675f4bcbfc-7v6b4\" (UID: \"999a6b3f-d0b8-477e-8be5-73488425a861\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.068636 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7v6b4"] Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.167614 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsvdr\" (UniqueName: \"kubernetes.io/projected/999a6b3f-d0b8-477e-8be5-73488425a861-kube-api-access-vsvdr\") pod \"dnsmasq-dns-675f4bcbfc-7v6b4\" (UID: \"999a6b3f-d0b8-477e-8be5-73488425a861\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.168180 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/999a6b3f-d0b8-477e-8be5-73488425a861-config\") pod \"dnsmasq-dns-675f4bcbfc-7v6b4\" (UID: \"999a6b3f-d0b8-477e-8be5-73488425a861\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.169852 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/999a6b3f-d0b8-477e-8be5-73488425a861-config\") pod \"dnsmasq-dns-675f4bcbfc-7v6b4\" (UID: \"999a6b3f-d0b8-477e-8be5-73488425a861\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.212030 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fwmgb"] Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.215394 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.217713 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsvdr\" (UniqueName: \"kubernetes.io/projected/999a6b3f-d0b8-477e-8be5-73488425a861-kube-api-access-vsvdr\") pod \"dnsmasq-dns-675f4bcbfc-7v6b4\" (UID: \"999a6b3f-d0b8-477e-8be5-73488425a861\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.224886 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fwmgb"] Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.225078 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.361367 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.382768 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hchh\" (UniqueName: \"kubernetes.io/projected/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-kube-api-access-5hchh\") pod \"dnsmasq-dns-78dd6ddcc-fwmgb\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.383015 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-config\") pod \"dnsmasq-dns-78dd6ddcc-fwmgb\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.383092 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-fwmgb\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.486340 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-config\") pod \"dnsmasq-dns-78dd6ddcc-fwmgb\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.486431 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-fwmgb\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.486617 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hchh\" (UniqueName: \"kubernetes.io/projected/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-kube-api-access-5hchh\") pod \"dnsmasq-dns-78dd6ddcc-fwmgb\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.488820 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-fwmgb\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.495405 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-config\") pod \"dnsmasq-dns-78dd6ddcc-fwmgb\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.516699 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hchh\" (UniqueName: \"kubernetes.io/projected/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-kube-api-access-5hchh\") pod \"dnsmasq-dns-78dd6ddcc-fwmgb\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 15:59:44 crc kubenswrapper[4767]: I0317 15:59:44.577615 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 15:59:45 crc kubenswrapper[4767]: I0317 15:59:45.069027 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7v6b4"] Mar 17 15:59:45 crc kubenswrapper[4767]: W0317 15:59:45.082755 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod999a6b3f_d0b8_477e_8be5_73488425a861.slice/crio-c1587499f610c4411e8f39dd2c401a96a14a494c75e04c184aa8d489ce404e8e WatchSource:0}: Error finding container c1587499f610c4411e8f39dd2c401a96a14a494c75e04c184aa8d489ce404e8e: Status 404 returned error can't find the container with id c1587499f610c4411e8f39dd2c401a96a14a494c75e04c184aa8d489ce404e8e Mar 17 15:59:45 crc kubenswrapper[4767]: I0317 15:59:45.424455 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fwmgb"] Mar 17 15:59:45 crc kubenswrapper[4767]: W0317 15:59:45.450041 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ae9edf0_7c73_4d24_a16c_04b6b3759a11.slice/crio-ad3ef96cafe9a7469b18d869a1cbccc040d8a4a18880aed4887f86d68a414ba3 WatchSource:0}: Error finding container ad3ef96cafe9a7469b18d869a1cbccc040d8a4a18880aed4887f86d68a414ba3: Status 404 returned error can't find the container with id ad3ef96cafe9a7469b18d869a1cbccc040d8a4a18880aed4887f86d68a414ba3 Mar 17 15:59:45 crc kubenswrapper[4767]: I0317 15:59:45.516659 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" event={"ID":"9ae9edf0-7c73-4d24-a16c-04b6b3759a11","Type":"ContainerStarted","Data":"ad3ef96cafe9a7469b18d869a1cbccc040d8a4a18880aed4887f86d68a414ba3"} Mar 17 15:59:45 crc kubenswrapper[4767]: I0317 15:59:45.518677 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" event={"ID":"999a6b3f-d0b8-477e-8be5-73488425a861","Type":"ContainerStarted","Data":"c1587499f610c4411e8f39dd2c401a96a14a494c75e04c184aa8d489ce404e8e"} Mar 17 15:59:46 crc kubenswrapper[4767]: I0317 15:59:46.563082 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7v6b4"] Mar 17 15:59:46 crc kubenswrapper[4767]: I0317 15:59:46.618372 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qzkct"] Mar 17 15:59:46 crc kubenswrapper[4767]: I0317 15:59:46.621153 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 15:59:46 crc kubenswrapper[4767]: I0317 15:59:46.634764 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qzkct"] Mar 17 15:59:46 crc kubenswrapper[4767]: I0317 15:59:46.798846 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5vqh\" (UniqueName: \"kubernetes.io/projected/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-kube-api-access-h5vqh\") pod \"dnsmasq-dns-666b6646f7-qzkct\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 15:59:46 crc kubenswrapper[4767]: I0317 15:59:46.799585 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-config\") pod \"dnsmasq-dns-666b6646f7-qzkct\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 15:59:46 crc kubenswrapper[4767]: I0317 15:59:46.799668 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qzkct\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.046737 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-config\") pod \"dnsmasq-dns-666b6646f7-qzkct\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.046830 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qzkct\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.046939 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5vqh\" (UniqueName: \"kubernetes.io/projected/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-kube-api-access-h5vqh\") pod \"dnsmasq-dns-666b6646f7-qzkct\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.048812 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-config\") pod \"dnsmasq-dns-666b6646f7-qzkct\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.048978 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qzkct\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.100078 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5vqh\" (UniqueName: \"kubernetes.io/projected/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-kube-api-access-h5vqh\") pod \"dnsmasq-dns-666b6646f7-qzkct\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.145458 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fwmgb"] Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.206102 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vvld"] Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.216440 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.250952 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5vvld\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.251030 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-config\") pod \"dnsmasq-dns-57d769cc4f-5vvld\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.251284 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj2f5\" (UniqueName: \"kubernetes.io/projected/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-kube-api-access-jj2f5\") pod \"dnsmasq-dns-57d769cc4f-5vvld\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.252557 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vvld"] Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.260034 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.562002 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj2f5\" (UniqueName: \"kubernetes.io/projected/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-kube-api-access-jj2f5\") pod \"dnsmasq-dns-57d769cc4f-5vvld\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.562261 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5vvld\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.562298 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-config\") pod \"dnsmasq-dns-57d769cc4f-5vvld\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.564474 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5vvld\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.595737 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-config\") pod \"dnsmasq-dns-57d769cc4f-5vvld\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.665014 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj2f5\" (UniqueName: \"kubernetes.io/projected/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-kube-api-access-jj2f5\") pod \"dnsmasq-dns-57d769cc4f-5vvld\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.753277 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.755537 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.761344 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.765382 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.765674 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.771530 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-p85jg" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.771805 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.771944 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.772059 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.774839 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.797284 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-1"] Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.799834 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.848830 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-2"] Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.870249 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.870541 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.887945 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.888002 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.889742 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-config-data\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.889784 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892100 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-pod-info\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892204 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892226 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5dp4\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-kube-api-access-x5dp4\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892264 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892316 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnqh9\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-kube-api-access-rnqh9\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892348 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-server-conf\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892367 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/048fa918-95b2-404e-81f5-8989ed47ec0a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892384 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/048fa918-95b2-404e-81f5-8989ed47ec0a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892407 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892492 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892512 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892541 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892571 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892625 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892652 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892680 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892716 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.892873 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-config-data\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.894466 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 15:59:47 crc kubenswrapper[4767]: I0317 15:59:47.917730 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000572 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-config-data\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000639 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-config-data\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000677 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000701 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000730 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000753 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-config-data\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000778 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000800 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000833 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-pod-info\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000849 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrjld\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-kube-api-access-lrjld\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000883 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000908 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5dp4\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-kube-api-access-x5dp4\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000938 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.000967 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnqh9\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-kube-api-access-rnqh9\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001006 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-server-conf\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001022 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-server-conf\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001041 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/048fa918-95b2-404e-81f5-8989ed47ec0a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001056 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/048fa918-95b2-404e-81f5-8989ed47ec0a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001072 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001097 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001121 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6bac37c2-c76e-45e2-82cf-78846509636f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001146 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001184 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001206 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001224 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001244 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001269 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001286 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-pod-info\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001305 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001325 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001340 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001360 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.001380 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.002973 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.003852 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-config-data\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.004477 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.004751 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.017634 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-server-conf\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.021967 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.025679 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.026549 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-pod-info\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.026944 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.027210 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.027272 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.027304 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/084c91f26dfb834887feef1465eec9c6367b6581a3e7afa9bcda1aeaba77086c/globalmount\"" pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.028027 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.028091 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8bdc0cdd497bb3d78a3d0a8594963d237808ef493d9599e21e04f965565170a3/globalmount\"" pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.031930 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.038855 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.042666 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.045084 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/048fa918-95b2-404e-81f5-8989ed47ec0a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.053679 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-config-data\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.056458 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.059149 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.062715 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnqh9\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-kube-api-access-rnqh9\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.067281 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5dp4\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-kube-api-access-x5dp4\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.082892 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/048fa918-95b2-404e-81f5-8989ed47ec0a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.109444 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.109570 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-config-data\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.109630 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.109678 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.109706 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrjld\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-kube-api-access-lrjld\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.109762 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-server-conf\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.109788 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.109810 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6bac37c2-c76e-45e2-82cf-78846509636f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.109845 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.109880 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-pod-info\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.109909 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.112552 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.112843 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.113548 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-config-data\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.121343 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.122980 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-server-conf\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.123662 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.141238 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.160084 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.161323 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.161376 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6bac37c2-c76e-45e2-82cf-78846509636f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/86d7f32ead6e09779828e242e22d0a054b17ff2cb152b126a42966414d82f3c6/globalmount\"" pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.181959 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-pod-info\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.186581 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrjld\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-kube-api-access-lrjld\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.747132 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.772602 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.981872 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 15:59:48 crc kubenswrapper[4767]: I0317 15:59:48.989610 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qzkct"] Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.010698 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\") pod \"rabbitmq-server-1\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " pod="openstack/rabbitmq-server-1" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.018809 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.018922 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.018939 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.019008 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-cczbq" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.018815 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.019148 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.019334 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 17 15:59:49 crc kubenswrapper[4767]: W0317 15:59:49.096906 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ff07a46_8cb7_4d77_ba6f_05012e4fa1d4.slice/crio-951d1fc99e8f8132d9ebc5baa315e0a8cf9dfcc34e2ba3724177cae2f99dc98f WatchSource:0}: Error finding container 951d1fc99e8f8132d9ebc5baa315e0a8cf9dfcc34e2ba3724177cae2f99dc98f: Status 404 returned error can't find the container with id 951d1fc99e8f8132d9ebc5baa315e0a8cf9dfcc34e2ba3724177cae2f99dc98f Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.151673 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") pod \"rabbitmq-server-0\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " pod="openstack/rabbitmq-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.196559 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.196620 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.196677 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.196774 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.196824 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.196859 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.196898 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.196920 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.196936 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.196974 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.196996 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nwj5\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-kube-api-access-8nwj5\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.296031 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.300518 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.300995 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.301100 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.301137 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.301389 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.301496 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.301560 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nwj5\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-kube-api-access-8nwj5\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.301606 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.301738 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.301821 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.302031 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.305013 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.315306 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.317723 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.324059 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.337129 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.337393 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.337427 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ceeeb535379e2ca1a9b6c52455f69317dc04eb485e1665f5be586667ac13ba46/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.347557 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.356052 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.356621 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6bac37c2-c76e-45e2-82cf-78846509636f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f\") pod \"rabbitmq-server-2\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " pod="openstack/rabbitmq-server-2" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.360558 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.364794 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.377071 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nwj5\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-kube-api-access-8nwj5\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.409095 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.826924 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 17 15:59:49 crc kubenswrapper[4767]: I0317 15:59:49.900057 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vvld"] Mar 17 15:59:50 crc kubenswrapper[4767]: I0317 15:59:50.133947 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:50 crc kubenswrapper[4767]: W0317 15:59:50.240953 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fa7ad38_b142_4986_a750_3e7c91d9fbb7.slice/crio-771d0422bbcf10179f8ba39dd37d42fe9c9ae847b713618ae34867d3e27b4fb0 WatchSource:0}: Error finding container 771d0422bbcf10179f8ba39dd37d42fe9c9ae847b713618ae34867d3e27b4fb0: Status 404 returned error can't find the container with id 771d0422bbcf10179f8ba39dd37d42fe9c9ae847b713618ae34867d3e27b4fb0 Mar 17 15:59:50 crc kubenswrapper[4767]: I0317 15:59:50.247043 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qzkct" event={"ID":"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4","Type":"ContainerStarted","Data":"951d1fc99e8f8132d9ebc5baa315e0a8cf9dfcc34e2ba3724177cae2f99dc98f"} Mar 17 15:59:50 crc kubenswrapper[4767]: I0317 15:59:50.312779 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-cczbq" Mar 17 15:59:50 crc kubenswrapper[4767]: I0317 15:59:50.324758 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.236581 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.335821 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.341711 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.349123 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.358276 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.359114 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-gblb6" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.360480 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.368130 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.457807 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hdtl\" (UniqueName: \"kubernetes.io/projected/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-kube-api-access-2hdtl\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.458503 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.458559 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.458597 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.458631 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7eead818-05b9-4016-b9ec-c2285f2d0fc2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7eead818-05b9-4016-b9ec-c2285f2d0fc2\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.458722 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-config-data-default\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.458808 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.458846 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-kolla-config\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.480060 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.778816 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hdtl\" (UniqueName: \"kubernetes.io/projected/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-kube-api-access-2hdtl\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.778887 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.778926 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.778955 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.778976 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7eead818-05b9-4016-b9ec-c2285f2d0fc2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7eead818-05b9-4016-b9ec-c2285f2d0fc2\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.779043 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-config-data-default\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.779096 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.779122 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-kolla-config\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.780457 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-kolla-config\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.786467 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" event={"ID":"8fa7ad38-b142-4986-a750-3e7c91d9fbb7","Type":"ContainerStarted","Data":"771d0422bbcf10179f8ba39dd37d42fe9c9ae847b713618ae34867d3e27b4fb0"} Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.792477 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-config-data-default\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.795380 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.797122 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.802758 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.803007 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.808719 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.852098 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.852946 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7eead818-05b9-4016-b9ec-c2285f2d0fc2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7eead818-05b9-4016-b9ec-c2285f2d0fc2\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/297718857a5815c6f0dd5c884a64b5950e5365dc0ce7ba240624877141866b50/globalmount\"" pod="openstack/openstack-galera-0" Mar 17 15:59:51 crc kubenswrapper[4767]: I0317 15:59:51.910713 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hdtl\" (UniqueName: \"kubernetes.io/projected/b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271-kube-api-access-2hdtl\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.008724 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7eead818-05b9-4016-b9ec-c2285f2d0fc2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7eead818-05b9-4016-b9ec-c2285f2d0fc2\") pod \"openstack-galera-0\" (UID: \"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271\") " pod="openstack/openstack-galera-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.475871 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.589182 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.605909 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.642020 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.642839 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-g7vq9" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.651416 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.664262 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.702478 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fa1b7ae1-40bc-474b-9602-de66005bea1c-kolla-config\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.702537 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b7ae1-40bc-474b-9602-de66005bea1c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.702603 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa1b7ae1-40bc-474b-9602-de66005bea1c-config-data\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.702671 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7zxm\" (UniqueName: \"kubernetes.io/projected/fa1b7ae1-40bc-474b-9602-de66005bea1c-kube-api-access-z7zxm\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.702728 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1b7ae1-40bc-474b-9602-de66005bea1c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.804597 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1b7ae1-40bc-474b-9602-de66005bea1c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.804672 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fa1b7ae1-40bc-474b-9602-de66005bea1c-kolla-config\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.804696 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b7ae1-40bc-474b-9602-de66005bea1c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.804761 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa1b7ae1-40bc-474b-9602-de66005bea1c-config-data\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:52 crc kubenswrapper[4767]: I0317 15:59:52.804824 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7zxm\" (UniqueName: \"kubernetes.io/projected/fa1b7ae1-40bc-474b-9602-de66005bea1c-kube-api-access-z7zxm\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.076136 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fa1b7ae1-40bc-474b-9602-de66005bea1c-kolla-config\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.085085 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa1b7ae1-40bc-474b-9602-de66005bea1c-config-data\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.098806 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b7ae1-40bc-474b-9602-de66005bea1c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.106315 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1b7ae1-40bc-474b-9602-de66005bea1c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.137565 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.220291 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.239297 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7zxm\" (UniqueName: \"kubernetes.io/projected/fa1b7ae1-40bc-474b-9602-de66005bea1c-kube-api-access-z7zxm\") pod \"memcached-0\" (UID: \"fa1b7ae1-40bc-474b-9602-de66005bea1c\") " pod="openstack/memcached-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.245763 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"048fa918-95b2-404e-81f5-8989ed47ec0a","Type":"ContainerStarted","Data":"4ff0b4a1c146a1bbe72ca66e23a27c07c1b21b664e2f47a9b254a15aa26541d4"} Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.257412 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"3ad3d357-572b-4b20-bfe6-4aa0b3513d41","Type":"ContainerStarted","Data":"a84ed7d7d41f56cd90679c1b7794d35f967406fcc5a89dac0435df1d24fe42cb"} Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.262615 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.264834 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.272372 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-xthds" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.272578 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.278062 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.278468 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.290011 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.314945 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.399065 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6abe028c-416e-4978-bcc8-3a7b8d92624b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.400243 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6abe028c-416e-4978-bcc8-3a7b8d92624b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.400360 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dghx8\" (UniqueName: \"kubernetes.io/projected/6abe028c-416e-4978-bcc8-3a7b8d92624b-kube-api-access-dghx8\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.400518 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6abe028c-416e-4978-bcc8-3a7b8d92624b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.409444 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6abe028c-416e-4978-bcc8-3a7b8d92624b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.409664 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-128c7722-6673-409c-9b53-321a75bcc2dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-128c7722-6673-409c-9b53-321a75bcc2dc\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.409717 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6abe028c-416e-4978-bcc8-3a7b8d92624b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.409753 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6abe028c-416e-4978-bcc8-3a7b8d92624b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.511733 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6abe028c-416e-4978-bcc8-3a7b8d92624b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.511818 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6abe028c-416e-4978-bcc8-3a7b8d92624b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.511923 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-128c7722-6673-409c-9b53-321a75bcc2dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-128c7722-6673-409c-9b53-321a75bcc2dc\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.511948 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6abe028c-416e-4978-bcc8-3a7b8d92624b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.511993 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6abe028c-416e-4978-bcc8-3a7b8d92624b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.512105 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6abe028c-416e-4978-bcc8-3a7b8d92624b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.512198 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6abe028c-416e-4978-bcc8-3a7b8d92624b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.512267 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dghx8\" (UniqueName: \"kubernetes.io/projected/6abe028c-416e-4978-bcc8-3a7b8d92624b-kube-api-access-dghx8\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.516951 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6abe028c-416e-4978-bcc8-3a7b8d92624b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.517869 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6abe028c-416e-4978-bcc8-3a7b8d92624b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.521408 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6abe028c-416e-4978-bcc8-3a7b8d92624b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.524896 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6abe028c-416e-4978-bcc8-3a7b8d92624b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.539803 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.539869 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-128c7722-6673-409c-9b53-321a75bcc2dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-128c7722-6673-409c-9b53-321a75bcc2dc\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fa4260840b47c09f4c055fbd1b3014594f045f7552b560374c5ebfdaf9f28fff/globalmount\"" pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.552301 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6abe028c-416e-4978-bcc8-3a7b8d92624b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.561542 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6abe028c-416e-4978-bcc8-3a7b8d92624b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.575780 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dghx8\" (UniqueName: \"kubernetes.io/projected/6abe028c-416e-4978-bcc8-3a7b8d92624b-kube-api-access-dghx8\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.764533 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-128c7722-6673-409c-9b53-321a75bcc2dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-128c7722-6673-409c-9b53-321a75bcc2dc\") pod \"openstack-cell1-galera-0\" (UID: \"6abe028c-416e-4978-bcc8-3a7b8d92624b\") " pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:53 crc kubenswrapper[4767]: I0317 15:59:53.928551 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 17 15:59:54 crc kubenswrapper[4767]: I0317 15:59:54.017812 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 17 15:59:54 crc kubenswrapper[4767]: I0317 15:59:54.414672 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e","Type":"ContainerStarted","Data":"3f40325b824810ca3a4f2de9038858fc4da068651267cf17d245c9681d1b797c"} Mar 17 15:59:54 crc kubenswrapper[4767]: I0317 15:59:54.434258 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271","Type":"ContainerStarted","Data":"a51529a135c878e9cfa83d31d64972d80de2bea9575e6c20e908fd8a5153b155"} Mar 17 15:59:54 crc kubenswrapper[4767]: I0317 15:59:54.481308 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a082b2bf-8b94-40e7-be0c-be64f75a4c3e","Type":"ContainerStarted","Data":"e8c41589e69f8552b3017809e34206b546ad0edfaaa82bfd1e7e2322f1b3691f"} Mar 17 15:59:54 crc kubenswrapper[4767]: I0317 15:59:54.628996 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 17 15:59:55 crc kubenswrapper[4767]: I0317 15:59:55.964683 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" podUID="7381ff68-5e5d-4281-9924-8495ece760f4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.102594 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" podUID="ad853953-ec48-40fc-8787-b2b838c955e9" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.245661 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" podUID="e59ff072-51c2-4995-9f61-709f12a1393a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.119810 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" podUID="59ba3b25-1e95-41ad-921c-9ee4ec5e2c43" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.246047 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" podUID="e59ff072-51c2-4995-9f61-709f12a1393a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.160873 4767 patch_prober.go:28] interesting pod/monitoring-plugin-6bbd5f6758-kwcw5 container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.85:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.246106 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" podUID="93033ac4-fd81-46eb-8014-184056ab6de2" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.85:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.246849 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" podUID="ad853953-ec48-40fc-8787-b2b838c955e9" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.116437 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" podUID="7381ff68-5e5d-4281-9924-8495ece760f4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.656465 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:8081/ready\": context deadline exceeded" start-of-body= Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.656602 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.56:8081/ready\": context deadline exceeded" Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.670094 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-8q5h4 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8081/ready\": context deadline exceeded" start-of-body= Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.670319 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podUID="560edfa4-8a88-4c9b-8b31-e61f93050c15" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/ready\": context deadline exceeded" Mar 17 15:59:56 crc kubenswrapper[4767]: E0317 15:59:56.742673 4767 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.254s" Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.742764 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fa1b7ae1-40bc-474b-9602-de66005bea1c","Type":"ContainerStarted","Data":"52073565ee099f91dd2d5c47bed3898078343b410ec6b8b37240a8625f0232f6"} Mar 17 15:59:56 crc kubenswrapper[4767]: I0317 15:59:56.742949 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 17 15:59:58 crc kubenswrapper[4767]: I0317 15:59:58.130529 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6abe028c-416e-4978-bcc8-3a7b8d92624b","Type":"ContainerStarted","Data":"257a18cb8c97dccb0a1533a15634ff38105f7945806a140463ed6eb1d373c9fa"} Mar 17 15:59:58 crc kubenswrapper[4767]: I0317 15:59:58.789258 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 15:59:58 crc kubenswrapper[4767]: I0317 15:59:58.791149 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 15:59:58 crc kubenswrapper[4767]: I0317 15:59:58.808867 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-kflhb" Mar 17 15:59:58 crc kubenswrapper[4767]: I0317 15:59:58.811073 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 15:59:58 crc kubenswrapper[4767]: I0317 15:59:58.826909 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knd9k\" (UniqueName: \"kubernetes.io/projected/232c4872-a09a-4be2-a7d2-c3c7232bae8a-kube-api-access-knd9k\") pod \"kube-state-metrics-0\" (UID: \"232c4872-a09a-4be2-a7d2-c3c7232bae8a\") " pod="openstack/kube-state-metrics-0" Mar 17 15:59:58 crc kubenswrapper[4767]: I0317 15:59:58.937709 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knd9k\" (UniqueName: \"kubernetes.io/projected/232c4872-a09a-4be2-a7d2-c3c7232bae8a-kube-api-access-knd9k\") pod \"kube-state-metrics-0\" (UID: \"232c4872-a09a-4be2-a7d2-c3c7232bae8a\") " pod="openstack/kube-state-metrics-0" Mar 17 15:59:59 crc kubenswrapper[4767]: I0317 15:59:59.245844 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knd9k\" (UniqueName: \"kubernetes.io/projected/232c4872-a09a-4be2-a7d2-c3c7232bae8a-kube-api-access-knd9k\") pod \"kube-state-metrics-0\" (UID: \"232c4872-a09a-4be2-a7d2-c3c7232bae8a\") " pod="openstack/kube-state-metrics-0" Mar 17 15:59:59 crc kubenswrapper[4767]: I0317 15:59:59.460202 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.248910 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562720-bgk8g"] Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.251553 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562720-bgk8g" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.270479 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.506542 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p2sd\" (UniqueName: \"kubernetes.io/projected/93755875-2f62-4a18-aa46-a2c55496e1c2-kube-api-access-4p2sd\") pod \"auto-csr-approver-29562720-bgk8g\" (UID: \"93755875-2f62-4a18-aa46-a2c55496e1c2\") " pod="openshift-infra/auto-csr-approver-29562720-bgk8g" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.512726 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82"] Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.515249 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.531138 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-47tcm"] Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.533795 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.543769 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82"] Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.563076 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-kqr5p"] Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.567720 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.570054 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.570429 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.570895 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.571800 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.572059 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.572261 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-g27hm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.572797 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.575600 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562720-bgk8g"] Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.590956 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-47tcm"] Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.610133 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-var-log-ovn\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.641154 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4lpp\" (UniqueName: \"kubernetes.io/projected/2aa82829-5daf-458e-b90d-705edcbbd3f2-kube-api-access-f4lpp\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.641312 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-ovn-controller-tls-certs\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.641396 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p2sd\" (UniqueName: \"kubernetes.io/projected/93755875-2f62-4a18-aa46-a2c55496e1c2-kube-api-access-4p2sd\") pod \"auto-csr-approver-29562720-bgk8g\" (UID: \"93755875-2f62-4a18-aa46-a2c55496e1c2\") " pod="openshift-infra/auto-csr-approver-29562720-bgk8g" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.640413 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-kqr5p"] Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.650008 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-config-volume\") pod \"collect-profiles-29562720-54h82\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.650071 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-var-run-ovn\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.650148 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-var-log\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.650191 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-var-run\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.650376 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-combined-ca-bundle\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.653443 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gthwq\" (UniqueName: \"kubernetes.io/projected/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-kube-api-access-gthwq\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.653622 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-var-lib\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.653687 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-var-run\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.653751 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-secret-volume\") pod \"collect-profiles-29562720-54h82\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.653813 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-etc-ovs\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.653845 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-scripts\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.654049 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmtdf\" (UniqueName: \"kubernetes.io/projected/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-kube-api-access-pmtdf\") pod \"collect-profiles-29562720-54h82\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.654118 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2aa82829-5daf-458e-b90d-705edcbbd3f2-scripts\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.767465 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gthwq\" (UniqueName: \"kubernetes.io/projected/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-kube-api-access-gthwq\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.767623 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-var-lib\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.767711 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-var-run\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.767797 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-secret-volume\") pod \"collect-profiles-29562720-54h82\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.767889 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-etc-ovs\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.767968 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-scripts\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.768082 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmtdf\" (UniqueName: \"kubernetes.io/projected/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-kube-api-access-pmtdf\") pod \"collect-profiles-29562720-54h82\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.768135 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2aa82829-5daf-458e-b90d-705edcbbd3f2-scripts\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.768235 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-var-log-ovn\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.768277 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4lpp\" (UniqueName: \"kubernetes.io/projected/2aa82829-5daf-458e-b90d-705edcbbd3f2-kube-api-access-f4lpp\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.768316 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-ovn-controller-tls-certs\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.768409 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-config-volume\") pod \"collect-profiles-29562720-54h82\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.768445 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-var-run-ovn\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.768504 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-var-log\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.768541 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-var-run\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.768609 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-combined-ca-bundle\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.935454 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2aa82829-5daf-458e-b90d-705edcbbd3f2-scripts\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.936625 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-var-lib\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.936826 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-var-run\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.943805 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-secret-volume\") pod \"collect-profiles-29562720-54h82\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.944030 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-etc-ovs\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.945945 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-scripts\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.946562 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-var-log-ovn\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.946818 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2aa82829-5daf-458e-b90d-705edcbbd3f2-var-log\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.946900 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-var-run-ovn\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.946941 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-var-run\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.948038 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-config-volume\") pod \"collect-profiles-29562720-54h82\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.951845 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-ovn-controller-tls-certs\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.966860 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p2sd\" (UniqueName: \"kubernetes.io/projected/93755875-2f62-4a18-aa46-a2c55496e1c2-kube-api-access-4p2sd\") pod \"auto-csr-approver-29562720-bgk8g\" (UID: \"93755875-2f62-4a18-aa46-a2c55496e1c2\") " pod="openshift-infra/auto-csr-approver-29562720-bgk8g" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.971847 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-combined-ca-bundle\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:00 crc kubenswrapper[4767]: I0317 16:00:00.988818 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562720-bgk8g" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.021320 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f"] Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.049271 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.050508 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4lpp\" (UniqueName: \"kubernetes.io/projected/2aa82829-5daf-458e-b90d-705edcbbd3f2-kube-api-access-f4lpp\") pod \"ovn-controller-ovs-kqr5p\" (UID: \"2aa82829-5daf-458e-b90d-705edcbbd3f2\") " pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.067701 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.085220 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-h4s59" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.102926 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pthwb\" (UniqueName: \"kubernetes.io/projected/9cd12dac-a828-4b81-bb16-723a38cf048d-kube-api-access-pthwb\") pod \"observability-ui-dashboards-7f87b9b85b-6h96f\" (UID: \"9cd12dac-a828-4b81-bb16-723a38cf048d\") " pod="openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.106336 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9cd12dac-a828-4b81-bb16-723a38cf048d-serving-cert\") pod \"observability-ui-dashboards-7f87b9b85b-6h96f\" (UID: \"9cd12dac-a828-4b81-bb16-723a38cf048d\") " pod="openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.122633 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f"] Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.149662 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmtdf\" (UniqueName: \"kubernetes.io/projected/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-kube-api-access-pmtdf\") pod \"collect-profiles-29562720-54h82\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.159194 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.166356 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gthwq\" (UniqueName: \"kubernetes.io/projected/d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0-kube-api-access-gthwq\") pod \"ovn-controller-47tcm\" (UID: \"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0\") " pod="openstack/ovn-controller-47tcm" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.214282 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pthwb\" (UniqueName: \"kubernetes.io/projected/9cd12dac-a828-4b81-bb16-723a38cf048d-kube-api-access-pthwb\") pod \"observability-ui-dashboards-7f87b9b85b-6h96f\" (UID: \"9cd12dac-a828-4b81-bb16-723a38cf048d\") " pod="openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.214388 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9cd12dac-a828-4b81-bb16-723a38cf048d-serving-cert\") pod \"observability-ui-dashboards-7f87b9b85b-6h96f\" (UID: \"9cd12dac-a828-4b81-bb16-723a38cf048d\") " pod="openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.235275 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.246049 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9cd12dac-a828-4b81-bb16-723a38cf048d-serving-cert\") pod \"observability-ui-dashboards-7f87b9b85b-6h96f\" (UID: \"9cd12dac-a828-4b81-bb16-723a38cf048d\") " pod="openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.247892 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.272962 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pthwb\" (UniqueName: \"kubernetes.io/projected/9cd12dac-a828-4b81-bb16-723a38cf048d-kube-api-access-pthwb\") pod \"observability-ui-dashboards-7f87b9b85b-6h96f\" (UID: \"9cd12dac-a828-4b81-bb16-723a38cf048d\") " pod="openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.272991 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.273314 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.273722 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.273786 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.274240 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-97b2h" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.275681 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.276014 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.296584 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.298633 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.357307 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.423244 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-47tcm" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.437393 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.437516 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.437597 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sbfg\" (UniqueName: \"kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-kube-api-access-4sbfg\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.437678 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.437706 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.437764 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.437798 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.437842 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.437907 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.438093 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.557255 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.557812 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.557902 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.557964 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sbfg\" (UniqueName: \"kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-kube-api-access-4sbfg\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.558060 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.558259 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.562617 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.572726 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.572916 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.572951 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.573003 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.573082 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.575348 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.576101 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.588750 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.600596 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.617136 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.622736 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.627924 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sbfg\" (UniqueName: \"kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-kube-api-access-4sbfg\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.665820 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.691311 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-89866dfb6-fswnc"] Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.694043 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.811014 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:00:01 crc kubenswrapper[4767]: I0317 16:00:01.811590 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0b502474e72a7e0212a814002fbab995e6f4dc4694e1fb66ec323aa24c386d38/globalmount\"" pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.059034 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-console-serving-cert\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.059253 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-service-ca\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.059441 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-oauth-serving-cert\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.059592 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-console-oauth-config\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.059656 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-console-config\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.059746 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5btx2\" (UniqueName: \"kubernetes.io/projected/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-kube-api-access-5btx2\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.059802 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-trusted-ca-bundle\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.075807 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-89866dfb6-fswnc"] Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.167074 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-oauth-serving-cert\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.167238 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-console-oauth-config\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.167280 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-console-config\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.167348 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5btx2\" (UniqueName: \"kubernetes.io/projected/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-kube-api-access-5btx2\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.167375 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-trusted-ca-bundle\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.167465 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-console-serving-cert\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.167539 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-service-ca\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.169476 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-service-ca\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.170356 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-console-config\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.170949 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-trusted-ca-bundle\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.171114 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-oauth-serving-cert\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.214148 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-console-serving-cert\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.217988 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-console-oauth-config\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.238712 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5btx2\" (UniqueName: \"kubernetes.io/projected/b52b7ad3-98cb-4051-8e3b-665fa44b0fd6-kube-api-access-5btx2\") pod \"console-89866dfb6-fswnc\" (UID: \"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6\") " pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.361687 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\") pod \"prometheus-metric-storage-0\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.392229 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.552569 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 17 16:00:02 crc kubenswrapper[4767]: I0317 16:00:02.678101 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 16:00:02 crc kubenswrapper[4767]: W0317 16:00:02.979327 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod232c4872_a09a_4be2_a7d2_c3c7232bae8a.slice/crio-b6284e8bcfda5776b5b64b781869d4c02c978fce8f3fc85c85d92be08c33d037 WatchSource:0}: Error finding container b6284e8bcfda5776b5b64b781869d4c02c978fce8f3fc85c85d92be08c33d037: Status 404 returned error can't find the container with id b6284e8bcfda5776b5b64b781869d4c02c978fce8f3fc85c85d92be08c33d037 Mar 17 16:00:03 crc kubenswrapper[4767]: I0317 16:00:03.995121 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"232c4872-a09a-4be2-a7d2-c3c7232bae8a","Type":"ContainerStarted","Data":"b6284e8bcfda5776b5b64b781869d4c02c978fce8f3fc85c85d92be08c33d037"} Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.018909 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562720-bgk8g"] Mar 17 16:00:04 crc kubenswrapper[4767]: W0317 16:00:04.032268 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93755875_2f62_4a18_aa46_a2c55496e1c2.slice/crio-18c291a3f5a129afdd5ef20b479487cb0514533d490afccf556dabb91fdb12fb WatchSource:0}: Error finding container 18c291a3f5a129afdd5ef20b479487cb0514533d490afccf556dabb91fdb12fb: Status 404 returned error can't find the container with id 18c291a3f5a129afdd5ef20b479487cb0514533d490afccf556dabb91fdb12fb Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.183243 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.183342 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.271800 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-47tcm"] Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.656843 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.699244 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.710424 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-z7k6w" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.710793 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.711106 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.716607 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.721980 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.792995 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.856917 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.865217 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a122d32-81b0-48a8-a327-06be6bb37899-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.865293 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a122d32-81b0-48a8-a327-06be6bb37899-config\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.865407 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a122d32-81b0-48a8-a327-06be6bb37899-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.865459 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a122d32-81b0-48a8-a327-06be6bb37899-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.865483 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a122d32-81b0-48a8-a327-06be6bb37899-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.865615 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpjkp\" (UniqueName: \"kubernetes.io/projected/8a122d32-81b0-48a8-a327-06be6bb37899-kube-api-access-zpjkp\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.865640 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8a122d32-81b0-48a8-a327-06be6bb37899-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.865713 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-913b97a9-d93a-4848-8d2c-ce1ba4908ba6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-913b97a9-d93a-4848-8d2c-ce1ba4908ba6\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.896017 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.912779 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.913121 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.913318 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.913792 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-ls56v" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.936396 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.968261 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a122d32-81b0-48a8-a327-06be6bb37899-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.968335 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a122d32-81b0-48a8-a327-06be6bb37899-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.968371 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a122d32-81b0-48a8-a327-06be6bb37899-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.968455 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpjkp\" (UniqueName: \"kubernetes.io/projected/8a122d32-81b0-48a8-a327-06be6bb37899-kube-api-access-zpjkp\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.968473 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8a122d32-81b0-48a8-a327-06be6bb37899-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.968529 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-913b97a9-d93a-4848-8d2c-ce1ba4908ba6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-913b97a9-d93a-4848-8d2c-ce1ba4908ba6\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.968593 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a122d32-81b0-48a8-a327-06be6bb37899-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.968616 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a122d32-81b0-48a8-a327-06be6bb37899-config\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.969975 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a122d32-81b0-48a8-a327-06be6bb37899-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.970510 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8a122d32-81b0-48a8-a327-06be6bb37899-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:04 crc kubenswrapper[4767]: I0317 16:00:04.988158 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a122d32-81b0-48a8-a327-06be6bb37899-config\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.014191 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a122d32-81b0-48a8-a327-06be6bb37899-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.015611 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a122d32-81b0-48a8-a327-06be6bb37899-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.015901 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a122d32-81b0-48a8-a327-06be6bb37899-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.034293 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpjkp\" (UniqueName: \"kubernetes.io/projected/8a122d32-81b0-48a8-a327-06be6bb37899-kube-api-access-zpjkp\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.065056 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82"] Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.084376 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ac6416e2-0457-4456-9cfe-0901b83f7b9b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac6416e2-0457-4456-9cfe-0901b83f7b9b\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.084864 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-config\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.085210 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.085560 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.085657 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.085762 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.085818 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.085964 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl4q2\" (UniqueName: \"kubernetes.io/projected/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-kube-api-access-vl4q2\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.098730 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.098788 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-913b97a9-d93a-4848-8d2c-ce1ba4908ba6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-913b97a9-d93a-4848-8d2c-ce1ba4908ba6\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f7155f2c35aac903982a2e0c741065d5c8df3279af84efa5cf45d397f2b0bcd1/globalmount\"" pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.131884 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" event={"ID":"62b3a4d3-4081-44b9-8d4b-fcebe2f89762","Type":"ContainerStarted","Data":"f8b14a521bdec248ab3a571693eb59622624f7e624b23e3e148e8c3566b8eee1"} Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.150390 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562720-bgk8g" event={"ID":"93755875-2f62-4a18-aa46-a2c55496e1c2","Type":"ContainerStarted","Data":"18c291a3f5a129afdd5ef20b479487cb0514533d490afccf556dabb91fdb12fb"} Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.171891 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-47tcm" event={"ID":"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0","Type":"ContainerStarted","Data":"6f512a9e1ea48bc17fbd5e3054cc82425ef0839cb5a480dc72225b1eaa872482"} Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.193676 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-config\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.193843 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.194015 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.194057 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.194455 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.194509 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.194609 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl4q2\" (UniqueName: \"kubernetes.io/projected/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-kube-api-access-vl4q2\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.194707 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ac6416e2-0457-4456-9cfe-0901b83f7b9b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac6416e2-0457-4456-9cfe-0901b83f7b9b\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.201662 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.203024 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.208378 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-config\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.209521 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.209585 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ac6416e2-0457-4456-9cfe-0901b83f7b9b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac6416e2-0457-4456-9cfe-0901b83f7b9b\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8fb7876bfef4e2f8356610ec2a383eabe4a127d24e8ef71a49836daa8a8010f9/globalmount\"" pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.224201 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.268594 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.270966 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl4q2\" (UniqueName: \"kubernetes.io/projected/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-kube-api-access-vl4q2\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.271689 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e7e6999-cfc1-4891-bbb4-48e848d2d8ef-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.362748 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f"] Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.400637 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-913b97a9-d93a-4848-8d2c-ce1ba4908ba6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-913b97a9-d93a-4848-8d2c-ce1ba4908ba6\") pod \"ovsdbserver-nb-0\" (UID: \"8a122d32-81b0-48a8-a327-06be6bb37899\") " pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.420872 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ac6416e2-0457-4456-9cfe-0901b83f7b9b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac6416e2-0457-4456-9cfe-0901b83f7b9b\") pod \"ovsdbserver-sb-0\" (UID: \"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef\") " pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.465244 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-89866dfb6-fswnc"] Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.465312 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.768358 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.769666 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 17 16:00:05 crc kubenswrapper[4767]: W0317 16:00:05.873583 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb52b7ad3_98cb_4051_8e3b_665fa44b0fd6.slice/crio-7afae11f0b43ee602c0e90836862c250f7ecd3928646800dc0336ef827c76fdb WatchSource:0}: Error finding container 7afae11f0b43ee602c0e90836862c250f7ecd3928646800dc0336ef827c76fdb: Status 404 returned error can't find the container with id 7afae11f0b43ee602c0e90836862c250f7ecd3928646800dc0336ef827c76fdb Mar 17 16:00:05 crc kubenswrapper[4767]: I0317 16:00:05.875579 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-kqr5p"] Mar 17 16:00:06 crc kubenswrapper[4767]: I0317 16:00:06.353207 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kqr5p" event={"ID":"2aa82829-5daf-458e-b90d-705edcbbd3f2","Type":"ContainerStarted","Data":"7e6cbfadce58a48b3a9c6477807ee479c3ca690967a15b880ba40352b124ff58"} Mar 17 16:00:06 crc kubenswrapper[4767]: I0317 16:00:06.364428 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f" event={"ID":"9cd12dac-a828-4b81-bb16-723a38cf048d","Type":"ContainerStarted","Data":"62a2df3de15b4eb3b9bad4b734881104a78dd40cee90b33ca1dd2aadf116c79c"} Mar 17 16:00:06 crc kubenswrapper[4767]: I0317 16:00:06.376721 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-89866dfb6-fswnc" event={"ID":"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6","Type":"ContainerStarted","Data":"7afae11f0b43ee602c0e90836862c250f7ecd3928646800dc0336ef827c76fdb"} Mar 17 16:00:06 crc kubenswrapper[4767]: I0317 16:00:06.394553 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f","Type":"ContainerStarted","Data":"6ff9ea01d5c579f3323ef089293e6ffdc95b914337844d915782fa2e02bdcdb3"} Mar 17 16:00:12 crc kubenswrapper[4767]: I0317 16:00:11.573696 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 16:00:12 crc kubenswrapper[4767]: I0317 16:00:11.599979 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 16:00:12 crc kubenswrapper[4767]: I0317 16:00:11.580451 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: i/o timeout (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 16:00:12 crc kubenswrapper[4767]: I0317 16:00:11.618838 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: i/o timeout (Client.Timeout exceeded while awaiting headers)" Mar 17 16:00:12 crc kubenswrapper[4767]: I0317 16:00:11.583449 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-frl2d" podUID="201932aa-c539-434b-992e-92f889e52de3" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 16:00:12 crc kubenswrapper[4767]: I0317 16:00:11.761666 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 17 16:00:12 crc kubenswrapper[4767]: I0317 16:00:12.917260 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 17 16:00:13 crc kubenswrapper[4767]: W0317 16:00:13.696530 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a122d32_81b0_48a8_a327_06be6bb37899.slice/crio-2594aedcebb46f96f7a66d81e986deecf756342724ba1b773200dd3ea18eb49b WatchSource:0}: Error finding container 2594aedcebb46f96f7a66d81e986deecf756342724ba1b773200dd3ea18eb49b: Status 404 returned error can't find the container with id 2594aedcebb46f96f7a66d81e986deecf756342724ba1b773200dd3ea18eb49b Mar 17 16:00:13 crc kubenswrapper[4767]: I0317 16:00:13.865050 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef","Type":"ContainerStarted","Data":"d3208946271c79daf4920b490972ae2ff873297aa0b1999dd194cdefed12a155"} Mar 17 16:00:13 crc kubenswrapper[4767]: I0317 16:00:13.886559 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8a122d32-81b0-48a8-a327-06be6bb37899","Type":"ContainerStarted","Data":"2594aedcebb46f96f7a66d81e986deecf756342724ba1b773200dd3ea18eb49b"} Mar 17 16:00:14 crc kubenswrapper[4767]: I0317 16:00:14.957474 4767 generic.go:334] "Generic (PLEG): container finished" podID="62b3a4d3-4081-44b9-8d4b-fcebe2f89762" containerID="27ad7e5eaab2a567614bd41747169bac163317f973daeb5fdbdce15ecd3695e6" exitCode=0 Mar 17 16:00:14 crc kubenswrapper[4767]: I0317 16:00:14.957605 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" event={"ID":"62b3a4d3-4081-44b9-8d4b-fcebe2f89762","Type":"ContainerDied","Data":"27ad7e5eaab2a567614bd41747169bac163317f973daeb5fdbdce15ecd3695e6"} Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.113400 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.191474 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" event={"ID":"62b3a4d3-4081-44b9-8d4b-fcebe2f89762","Type":"ContainerDied","Data":"f8b14a521bdec248ab3a571693eb59622624f7e624b23e3e148e8c3566b8eee1"} Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.191577 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8b14a521bdec248ab3a571693eb59622624f7e624b23e3e148e8c3566b8eee1" Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.191637 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82" Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.216945 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-secret-volume\") pod \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.217262 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-config-volume\") pod \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.217384 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmtdf\" (UniqueName: \"kubernetes.io/projected/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-kube-api-access-pmtdf\") pod \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\" (UID: \"62b3a4d3-4081-44b9-8d4b-fcebe2f89762\") " Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.218590 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-config-volume" (OuterVolumeSpecName: "config-volume") pod "62b3a4d3-4081-44b9-8d4b-fcebe2f89762" (UID: "62b3a4d3-4081-44b9-8d4b-fcebe2f89762"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.224565 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-kube-api-access-pmtdf" (OuterVolumeSpecName: "kube-api-access-pmtdf") pod "62b3a4d3-4081-44b9-8d4b-fcebe2f89762" (UID: "62b3a4d3-4081-44b9-8d4b-fcebe2f89762"). InnerVolumeSpecName "kube-api-access-pmtdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.224968 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "62b3a4d3-4081-44b9-8d4b-fcebe2f89762" (UID: "62b3a4d3-4081-44b9-8d4b-fcebe2f89762"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.228818 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.228867 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmtdf\" (UniqueName: \"kubernetes.io/projected/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-kube-api-access-pmtdf\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:23 crc kubenswrapper[4767]: I0317 16:00:23.228882 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62b3a4d3-4081-44b9-8d4b-fcebe2f89762-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:25 crc kubenswrapper[4767]: I0317 16:00:25.262497 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-89866dfb6-fswnc" event={"ID":"b52b7ad3-98cb-4051-8e3b-665fa44b0fd6","Type":"ContainerStarted","Data":"ea50b8fe775e50ea6b2a5d1c1bfbcfc69a812578952afd7d8ce0b9aeba20041b"} Mar 17 16:00:25 crc kubenswrapper[4767]: I0317 16:00:25.290830 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-89866dfb6-fswnc" podStartSLOduration=24.290799091 podStartE2EDuration="24.290799091s" podCreationTimestamp="2026-03-17 16:00:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:00:25.286646653 +0000 UTC m=+1416.699962710" watchObservedRunningTime="2026-03-17 16:00:25.290799091 +0000 UTC m=+1416.704115138" Mar 17 16:00:27 crc kubenswrapper[4767]: E0317 16:00:27.896908 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Mar 17 16:00:27 crc kubenswrapper[4767]: E0317 16:00:27.897449 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8nwj5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(ef2226f0-7710-4a82-84ab-b4ce37f8bf2e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:27 crc kubenswrapper[4767]: E0317 16:00:27.898714 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" Mar 17 16:00:28 crc kubenswrapper[4767]: E0317 16:00:28.346483 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" Mar 17 16:00:32 crc kubenswrapper[4767]: I0317 16:00:32.394073 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:32 crc kubenswrapper[4767]: I0317 16:00:32.394636 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:32 crc kubenswrapper[4767]: I0317 16:00:32.400667 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:33 crc kubenswrapper[4767]: I0317 16:00:33.410922 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-89866dfb6-fswnc" Mar 17 16:00:33 crc kubenswrapper[4767]: I0317 16:00:33.602091 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5c7b759f69-p6dvt"] Mar 17 16:00:34 crc kubenswrapper[4767]: I0317 16:00:34.166577 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:00:34 crc kubenswrapper[4767]: I0317 16:00:34.166967 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:00:34 crc kubenswrapper[4767]: I0317 16:00:34.167034 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 16:00:34 crc kubenswrapper[4767]: I0317 16:00:34.167970 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"53b1541419163e416532d56412b5286cd823f9a0b2d0b4b97086b6642dba72af"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 16:00:34 crc kubenswrapper[4767]: I0317 16:00:34.168042 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://53b1541419163e416532d56412b5286cd823f9a0b2d0b4b97086b6642dba72af" gracePeriod=600 Mar 17 16:00:34 crc kubenswrapper[4767]: I0317 16:00:34.422364 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="53b1541419163e416532d56412b5286cd823f9a0b2d0b4b97086b6642dba72af" exitCode=0 Mar 17 16:00:34 crc kubenswrapper[4767]: I0317 16:00:34.422465 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"53b1541419163e416532d56412b5286cd823f9a0b2d0b4b97086b6642dba72af"} Mar 17 16:00:34 crc kubenswrapper[4767]: I0317 16:00:34.422922 4767 scope.go:117] "RemoveContainer" containerID="c0a59e1cd71c114bacc3d0ebc2a7a84e16351d4b20330e3bb7479cc322816c8b" Mar 17 16:00:36 crc kubenswrapper[4767]: E0317 16:00:36.421859 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Mar 17 16:00:36 crc kubenswrapper[4767]: E0317 16:00:36.422440 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rnqh9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(048fa918-95b2-404e-81f5-8989ed47ec0a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:36 crc kubenswrapper[4767]: E0317 16:00:36.424023 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="048fa918-95b2-404e-81f5-8989ed47ec0a" Mar 17 16:00:36 crc kubenswrapper[4767]: E0317 16:00:36.450152 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="048fa918-95b2-404e-81f5-8989ed47ec0a" Mar 17 16:00:38 crc kubenswrapper[4767]: E0317 16:00:38.873443 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Mar 17 16:00:38 crc kubenswrapper[4767]: E0317 16:00:38.873999 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2hdtl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:38 crc kubenswrapper[4767]: E0317 16:00:38.875358 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" Mar 17 16:00:38 crc kubenswrapper[4767]: E0317 16:00:38.882810 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Mar 17 16:00:38 crc kubenswrapper[4767]: E0317 16:00:38.883070 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x5dp4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-1_openstack(3ad3d357-572b-4b20-bfe6-4aa0b3513d41): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:38 crc kubenswrapper[4767]: E0317 16:00:38.884257 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-1" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" Mar 17 16:00:39 crc kubenswrapper[4767]: E0317 16:00:39.482837 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" Mar 17 16:00:39 crc kubenswrapper[4767]: E0317 16:00:39.482911 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-1" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.163632 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.164164 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:ovsdb-server-init,Image:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,Command:[/usr/local/bin/container-scripts/init-ovsdb-server.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n688h98h68fh547hbbh6bh577hfch56h5c4h5fch5cdh67chb8hcch75h5dbh669hfchd4h677h65dh589h57dh5d9h5c8h64dhc6h5f6h66dh55dh697q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-ovs,ReadOnly:false,MountPath:/etc/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log,ReadOnly:false,MountPath:/var/log/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-lib,ReadOnly:false,MountPath:/var/lib/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f4lpp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-ovs-kqr5p_openstack(2aa82829-5daf-458e-b90d-705edcbbd3f2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.180299 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.180758 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lrjld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-2_openstack(a082b2bf-8b94-40e7-be0c-be64f75a4c3e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.183086 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-2" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.196457 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-ovs-kqr5p" podUID="2aa82829-5daf-458e-b90d-705edcbbd3f2" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.222966 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.223249 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dghx8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(6abe028c-416e-4978-bcc8-3a7b8d92624b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.224557 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="6abe028c-416e-4978-bcc8-3a7b8d92624b" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.491611 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="6abe028c-416e-4978-bcc8-3a7b8d92624b" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.491686 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified\\\"\"" pod="openstack/ovn-controller-ovs-kqr5p" podUID="2aa82829-5daf-458e-b90d-705edcbbd3f2" Mar 17 16:00:40 crc kubenswrapper[4767]: E0317 16:00:40.492074 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-2" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" Mar 17 16:00:41 crc kubenswrapper[4767]: E0317 16:00:41.363288 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Mar 17 16:00:41 crc kubenswrapper[4767]: E0317 16:00:41.363504 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n5ffh55ch595h8ch58bh85h99h88h5bdhcdh7bh658h58ch87hc5h5d7h5f6h94h54fh9h598h78h9h4h64h5b5hdh5cfh544h689h68hf8q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z7zxm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(fa1b7ae1-40bc-474b-9602-de66005bea1c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:41 crc kubenswrapper[4767]: E0317 16:00:41.364678 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="fa1b7ae1-40bc-474b-9602-de66005bea1c" Mar 17 16:00:41 crc kubenswrapper[4767]: E0317 16:00:41.509029 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="fa1b7ae1-40bc-474b-9602-de66005bea1c" Mar 17 16:00:42 crc kubenswrapper[4767]: E0317 16:00:42.310298 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 17 16:00:42 crc kubenswrapper[4767]: E0317 16:00:42.310557 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jj2f5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-5vvld_openstack(8fa7ad38-b142-4986-a750-3e7c91d9fbb7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:42 crc kubenswrapper[4767]: E0317 16:00:42.312320 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" podUID="8fa7ad38-b142-4986-a750-3e7c91d9fbb7" Mar 17 16:00:42 crc kubenswrapper[4767]: E0317 16:00:42.342276 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 17 16:00:42 crc kubenswrapper[4767]: E0317 16:00:42.342540 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h5vqh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-qzkct_openstack(3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:42 crc kubenswrapper[4767]: E0317 16:00:42.366594 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-qzkct" podUID="3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4" Mar 17 16:00:42 crc kubenswrapper[4767]: E0317 16:00:42.450415 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 17 16:00:42 crc kubenswrapper[4767]: E0317 16:00:42.450639 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vsvdr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-7v6b4_openstack(999a6b3f-d0b8-477e-8be5-73488425a861): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:42 crc kubenswrapper[4767]: E0317 16:00:42.452023 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" podUID="999a6b3f-d0b8-477e-8be5-73488425a861" Mar 17 16:00:42 crc kubenswrapper[4767]: E0317 16:00:42.525464 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-qzkct" podUID="3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4" Mar 17 16:00:42 crc kubenswrapper[4767]: E0317 16:00:42.525727 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" podUID="8fa7ad38-b142-4986-a750-3e7c91d9fbb7" Mar 17 16:00:43 crc kubenswrapper[4767]: E0317 16:00:43.077490 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified" Mar 17 16:00:43 crc kubenswrapper[4767]: E0317 16:00:43.078231 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n688h98h68fh547hbbh6bh577hfch56h5c4h5fch5cdh67chb8hcch75h5dbh669hfchd4h677h65dh589h57dh5d9h5c8h64dhc6h5f6h66dh55dh697q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gthwq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-47tcm_openstack(d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:43 crc kubenswrapper[4767]: E0317 16:00:43.079469 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-47tcm" podUID="d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0" Mar 17 16:00:43 crc kubenswrapper[4767]: E0317 16:00:43.182634 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 17 16:00:43 crc kubenswrapper[4767]: E0317 16:00:43.182861 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5hchh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-fwmgb_openstack(9ae9edf0-7c73-4d24-a16c-04b6b3759a11): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:43 crc kubenswrapper[4767]: E0317 16:00:43.184006 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" podUID="9ae9edf0-7c73-4d24-a16c-04b6b3759a11" Mar 17 16:00:43 crc kubenswrapper[4767]: E0317 16:00:43.529693 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified\\\"\"" pod="openstack/ovn-controller-47tcm" podUID="d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.024313 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-42gbs"] Mar 17 16:00:45 crc kubenswrapper[4767]: E0317 16:00:45.025461 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b3a4d3-4081-44b9-8d4b-fcebe2f89762" containerName="collect-profiles" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.025483 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b3a4d3-4081-44b9-8d4b-fcebe2f89762" containerName="collect-profiles" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.025983 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b3a4d3-4081-44b9-8d4b-fcebe2f89762" containerName="collect-profiles" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.027391 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.029741 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.049665 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-42gbs"] Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.145328 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/12012344-284e-4ce3-949e-dcd3812587ee-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.145396 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12012344-284e-4ce3-949e-dcd3812587ee-combined-ca-bundle\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.145442 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff4v4\" (UniqueName: \"kubernetes.io/projected/12012344-284e-4ce3-949e-dcd3812587ee-kube-api-access-ff4v4\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.145491 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/12012344-284e-4ce3-949e-dcd3812587ee-ovs-rundir\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.145569 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/12012344-284e-4ce3-949e-dcd3812587ee-ovn-rundir\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.145728 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12012344-284e-4ce3-949e-dcd3812587ee-config\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.235398 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qzkct"] Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.247852 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12012344-284e-4ce3-949e-dcd3812587ee-config\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.247980 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/12012344-284e-4ce3-949e-dcd3812587ee-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.248006 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12012344-284e-4ce3-949e-dcd3812587ee-combined-ca-bundle\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.248072 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff4v4\" (UniqueName: \"kubernetes.io/projected/12012344-284e-4ce3-949e-dcd3812587ee-kube-api-access-ff4v4\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.248105 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/12012344-284e-4ce3-949e-dcd3812587ee-ovs-rundir\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.248164 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/12012344-284e-4ce3-949e-dcd3812587ee-ovn-rundir\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.248672 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/12012344-284e-4ce3-949e-dcd3812587ee-ovn-rundir\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.249087 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/12012344-284e-4ce3-949e-dcd3812587ee-ovs-rundir\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.249690 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12012344-284e-4ce3-949e-dcd3812587ee-config\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.255771 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/12012344-284e-4ce3-949e-dcd3812587ee-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.259955 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12012344-284e-4ce3-949e-dcd3812587ee-combined-ca-bundle\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.281423 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff4v4\" (UniqueName: \"kubernetes.io/projected/12012344-284e-4ce3-949e-dcd3812587ee-kube-api-access-ff4v4\") pod \"ovn-controller-metrics-42gbs\" (UID: \"12012344-284e-4ce3-949e-dcd3812587ee\") " pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.332625 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xhq4t"] Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.334767 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.355842 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.398598 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-42gbs" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.408761 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xhq4t"] Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.456966 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcr5w\" (UniqueName: \"kubernetes.io/projected/24aea215-34c0-40a4-a6fd-32a24222a50a-kube-api-access-dcr5w\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.457048 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.457098 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-config\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.457165 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.519615 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vvld"] Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.587480 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-zvjq6"] Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.590842 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.593705 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.593839 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcr5w\" (UniqueName: \"kubernetes.io/projected/24aea215-34c0-40a4-a6fd-32a24222a50a-kube-api-access-dcr5w\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.593894 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.593958 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-config\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.595718 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.598656 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-config\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.605563 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.611470 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.657373 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-zvjq6"] Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.700880 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.700967 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.711155 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcr5w\" (UniqueName: \"kubernetes.io/projected/24aea215-34c0-40a4-a6fd-32a24222a50a-kube-api-access-dcr5w\") pod \"dnsmasq-dns-7fd796d7df-xhq4t\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.727988 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-config\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.728057 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbxh9\" (UniqueName: \"kubernetes.io/projected/b2781552-ff04-4f51-8901-7abb87dd1369-kube-api-access-wbxh9\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.728139 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.836893 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-config\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.836965 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbxh9\" (UniqueName: \"kubernetes.io/projected/b2781552-ff04-4f51-8901-7abb87dd1369-kube-api-access-wbxh9\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.837017 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.837084 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.837114 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.837993 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.838579 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-config\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.841158 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.848929 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.879298 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbxh9\" (UniqueName: \"kubernetes.io/projected/b2781552-ff04-4f51-8901-7abb87dd1369-kube-api-access-wbxh9\") pod \"dnsmasq-dns-86db49b7ff-zvjq6\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:45 crc kubenswrapper[4767]: I0317 16:00:45.974874 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:46 crc kubenswrapper[4767]: I0317 16:00:46.020813 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:47 crc kubenswrapper[4767]: E0317 16:00:47.230311 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified" Mar 17 16:00:47 crc kubenswrapper[4767]: E0317 16:00:47.230629 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-sb,Image:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd7h668h5f5h664h5b8h5cfh688h9ch694h5d8h59dh686h57h568h668h5bh5ffh9dh5dch64fh55chb9h555h67bh546h65fhbdh89h675h695h669h65fq,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-sb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vl4q2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(4e7e6999-cfc1-4891-bbb4-48e848d2d8ef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:47 crc kubenswrapper[4767]: E0317 16:00:47.765611 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified" Mar 17 16:00:47 crc kubenswrapper[4767]: E0317 16:00:47.766325 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd5h57ch5f8h64bh55h644hdh6dh95hcfh656h598h59h5dch669h58h599h58bhb7h5bfh677h6chdh7ch57bh657h5bh5b8h577hbdh59ch56fq,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zpjkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(8a122d32-81b0-48a8-a327-06be6bb37899): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:00:47 crc kubenswrapper[4767]: I0317 16:00:47.997594 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" Mar 17 16:00:47 crc kubenswrapper[4767]: I0317 16:00:47.998432 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.015783 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.025418 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150081 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsvdr\" (UniqueName: \"kubernetes.io/projected/999a6b3f-d0b8-477e-8be5-73488425a861-kube-api-access-vsvdr\") pod \"999a6b3f-d0b8-477e-8be5-73488425a861\" (UID: \"999a6b3f-d0b8-477e-8be5-73488425a861\") " Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150145 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-dns-svc\") pod \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150205 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/999a6b3f-d0b8-477e-8be5-73488425a861-config\") pod \"999a6b3f-d0b8-477e-8be5-73488425a861\" (UID: \"999a6b3f-d0b8-477e-8be5-73488425a861\") " Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150240 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-dns-svc\") pod \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150286 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-config\") pod \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150425 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-config\") pod \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150488 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-config\") pod \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150580 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5vqh\" (UniqueName: \"kubernetes.io/projected/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-kube-api-access-h5vqh\") pod \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\" (UID: \"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4\") " Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150630 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-dns-svc\") pod \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150669 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj2f5\" (UniqueName: \"kubernetes.io/projected/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-kube-api-access-jj2f5\") pod \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\" (UID: \"8fa7ad38-b142-4986-a750-3e7c91d9fbb7\") " Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150777 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hchh\" (UniqueName: \"kubernetes.io/projected/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-kube-api-access-5hchh\") pod \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\" (UID: \"9ae9edf0-7c73-4d24-a16c-04b6b3759a11\") " Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.150901 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-config" (OuterVolumeSpecName: "config") pod "3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4" (UID: "3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.151793 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/999a6b3f-d0b8-477e-8be5-73488425a861-config" (OuterVolumeSpecName: "config") pod "999a6b3f-d0b8-477e-8be5-73488425a861" (UID: "999a6b3f-d0b8-477e-8be5-73488425a861"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.152560 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9ae9edf0-7c73-4d24-a16c-04b6b3759a11" (UID: "9ae9edf0-7c73-4d24-a16c-04b6b3759a11"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.153047 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-config" (OuterVolumeSpecName: "config") pod "8fa7ad38-b142-4986-a750-3e7c91d9fbb7" (UID: "8fa7ad38-b142-4986-a750-3e7c91d9fbb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.152984 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4" (UID: "3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.153408 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-config" (OuterVolumeSpecName: "config") pod "9ae9edf0-7c73-4d24-a16c-04b6b3759a11" (UID: "9ae9edf0-7c73-4d24-a16c-04b6b3759a11"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.153454 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8fa7ad38-b142-4986-a750-3e7c91d9fbb7" (UID: "8fa7ad38-b142-4986-a750-3e7c91d9fbb7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.154090 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.154111 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.154124 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.154137 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.154148 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/999a6b3f-d0b8-477e-8be5-73488425a861-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.154159 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.154187 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.159666 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/999a6b3f-d0b8-477e-8be5-73488425a861-kube-api-access-vsvdr" (OuterVolumeSpecName: "kube-api-access-vsvdr") pod "999a6b3f-d0b8-477e-8be5-73488425a861" (UID: "999a6b3f-d0b8-477e-8be5-73488425a861"). InnerVolumeSpecName "kube-api-access-vsvdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.159824 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-kube-api-access-h5vqh" (OuterVolumeSpecName: "kube-api-access-h5vqh") pod "3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4" (UID: "3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4"). InnerVolumeSpecName "kube-api-access-h5vqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.160093 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-kube-api-access-5hchh" (OuterVolumeSpecName: "kube-api-access-5hchh") pod "9ae9edf0-7c73-4d24-a16c-04b6b3759a11" (UID: "9ae9edf0-7c73-4d24-a16c-04b6b3759a11"). InnerVolumeSpecName "kube-api-access-5hchh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.171145 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-kube-api-access-jj2f5" (OuterVolumeSpecName: "kube-api-access-jj2f5") pod "8fa7ad38-b142-4986-a750-3e7c91d9fbb7" (UID: "8fa7ad38-b142-4986-a750-3e7c91d9fbb7"). InnerVolumeSpecName "kube-api-access-jj2f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.285860 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5vqh\" (UniqueName: \"kubernetes.io/projected/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4-kube-api-access-h5vqh\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.286215 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj2f5\" (UniqueName: \"kubernetes.io/projected/8fa7ad38-b142-4986-a750-3e7c91d9fbb7-kube-api-access-jj2f5\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.286236 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hchh\" (UniqueName: \"kubernetes.io/projected/9ae9edf0-7c73-4d24-a16c-04b6b3759a11-kube-api-access-5hchh\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.286249 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsvdr\" (UniqueName: \"kubernetes.io/projected/999a6b3f-d0b8-477e-8be5-73488425a861-kube-api-access-vsvdr\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.625732 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562720-bgk8g" event={"ID":"93755875-2f62-4a18-aa46-a2c55496e1c2","Type":"ContainerStarted","Data":"cc06b7dcb8534812e686a7f25518ad0d196b5c0386841ca0064c5eaba1dc1848"} Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.639397 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qzkct" event={"ID":"3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4","Type":"ContainerDied","Data":"951d1fc99e8f8132d9ebc5baa315e0a8cf9dfcc34e2ba3724177cae2f99dc98f"} Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.639543 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qzkct" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.647409 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" event={"ID":"9ae9edf0-7c73-4d24-a16c-04b6b3759a11","Type":"ContainerDied","Data":"ad3ef96cafe9a7469b18d869a1cbccc040d8a4a18880aed4887f86d68a414ba3"} Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.647567 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fwmgb" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.651260 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" event={"ID":"8fa7ad38-b142-4986-a750-3e7c91d9fbb7","Type":"ContainerDied","Data":"771d0422bbcf10179f8ba39dd37d42fe9c9ae847b713618ae34867d3e27b4fb0"} Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.651409 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vvld" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.670344 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" event={"ID":"999a6b3f-d0b8-477e-8be5-73488425a861","Type":"ContainerDied","Data":"c1587499f610c4411e8f39dd2c401a96a14a494c75e04c184aa8d489ce404e8e"} Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.670666 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7v6b4" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.672774 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562720-bgk8g" podStartSLOduration=29.319013707 podStartE2EDuration="48.672746075s" podCreationTimestamp="2026-03-17 16:00:00 +0000 UTC" firstStartedPulling="2026-03-17 16:00:04.115083275 +0000 UTC m=+1395.528399322" lastFinishedPulling="2026-03-17 16:00:23.468815643 +0000 UTC m=+1414.882131690" observedRunningTime="2026-03-17 16:00:48.663841067 +0000 UTC m=+1440.077157124" watchObservedRunningTime="2026-03-17 16:00:48.672746075 +0000 UTC m=+1440.086062122" Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.741728 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-42gbs"] Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.774972 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qzkct"] Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.802890 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qzkct"] Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.833281 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7v6b4"] Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.841909 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7v6b4"] Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.866441 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fwmgb"] Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.874446 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fwmgb"] Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.892479 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vvld"] Mar 17 16:00:48 crc kubenswrapper[4767]: I0317 16:00:48.900776 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vvld"] Mar 17 16:00:49 crc kubenswrapper[4767]: I0317 16:00:49.370644 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4" path="/var/lib/kubelet/pods/3ff07a46-8cb7-4d77-ba6f-05012e4fa1d4/volumes" Mar 17 16:00:49 crc kubenswrapper[4767]: I0317 16:00:49.371449 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fa7ad38-b142-4986-a750-3e7c91d9fbb7" path="/var/lib/kubelet/pods/8fa7ad38-b142-4986-a750-3e7c91d9fbb7/volumes" Mar 17 16:00:49 crc kubenswrapper[4767]: I0317 16:00:49.371909 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="999a6b3f-d0b8-477e-8be5-73488425a861" path="/var/lib/kubelet/pods/999a6b3f-d0b8-477e-8be5-73488425a861/volumes" Mar 17 16:00:49 crc kubenswrapper[4767]: I0317 16:00:49.372591 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ae9edf0-7c73-4d24-a16c-04b6b3759a11" path="/var/lib/kubelet/pods/9ae9edf0-7c73-4d24-a16c-04b6b3759a11/volumes" Mar 17 16:00:49 crc kubenswrapper[4767]: E0317 16:00:49.517762 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Mar 17 16:00:49 crc kubenswrapper[4767]: E0317 16:00:49.517833 4767 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Mar 17 16:00:49 crc kubenswrapper[4767]: E0317 16:00:49.517990 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-knd9k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(232c4872-a09a-4be2-a7d2-c3c7232bae8a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 16:00:49 crc kubenswrapper[4767]: E0317 16:00:49.519181 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="232c4872-a09a-4be2-a7d2-c3c7232bae8a" Mar 17 16:00:49 crc kubenswrapper[4767]: I0317 16:00:49.700372 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-42gbs" event={"ID":"12012344-284e-4ce3-949e-dcd3812587ee","Type":"ContainerStarted","Data":"9bdfc6866c2cab0a63807714c83f52d9dd50731560ca63aea8e4260ec5bd2e70"} Mar 17 16:00:49 crc kubenswrapper[4767]: I0317 16:00:49.702948 4767 generic.go:334] "Generic (PLEG): container finished" podID="93755875-2f62-4a18-aa46-a2c55496e1c2" containerID="cc06b7dcb8534812e686a7f25518ad0d196b5c0386841ca0064c5eaba1dc1848" exitCode=0 Mar 17 16:00:49 crc kubenswrapper[4767]: I0317 16:00:49.703019 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562720-bgk8g" event={"ID":"93755875-2f62-4a18-aa46-a2c55496e1c2","Type":"ContainerDied","Data":"cc06b7dcb8534812e686a7f25518ad0d196b5c0386841ca0064c5eaba1dc1848"} Mar 17 16:00:49 crc kubenswrapper[4767]: I0317 16:00:49.731651 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c"} Mar 17 16:00:49 crc kubenswrapper[4767]: E0317 16:00:49.760393 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="232c4872-a09a-4be2-a7d2-c3c7232bae8a" Mar 17 16:00:50 crc kubenswrapper[4767]: I0317 16:00:50.112051 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-zvjq6"] Mar 17 16:00:50 crc kubenswrapper[4767]: I0317 16:00:50.250671 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xhq4t"] Mar 17 16:00:50 crc kubenswrapper[4767]: I0317 16:00:50.744346 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f" event={"ID":"9cd12dac-a828-4b81-bb16-723a38cf048d","Type":"ContainerStarted","Data":"36cb75d1f31a9c75c2fcb4d972dd2dae56773bbd977e6619e3f5505a8e48a06e"} Mar 17 16:00:50 crc kubenswrapper[4767]: I0317 16:00:50.747524 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" event={"ID":"b2781552-ff04-4f51-8901-7abb87dd1369","Type":"ContainerStarted","Data":"6d2594ab02f6bf4e993b5fb971d163b9bcb627d76ae0805eee84fd9b153e650f"} Mar 17 16:00:50 crc kubenswrapper[4767]: I0317 16:00:50.749285 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" event={"ID":"24aea215-34c0-40a4-a6fd-32a24222a50a","Type":"ContainerStarted","Data":"cdbc9bea367092e702d46ead0f629d6c4fbf64a65f246ac40495d6826ce18e70"} Mar 17 16:00:50 crc kubenswrapper[4767]: I0317 16:00:50.767964 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7f87b9b85b-6h96f" podStartSLOduration=14.459729601 podStartE2EDuration="50.767939814s" podCreationTimestamp="2026-03-17 16:00:00 +0000 UTC" firstStartedPulling="2026-03-17 16:00:05.953110687 +0000 UTC m=+1397.366426734" lastFinishedPulling="2026-03-17 16:00:42.2613209 +0000 UTC m=+1433.674636947" observedRunningTime="2026-03-17 16:00:50.763254759 +0000 UTC m=+1442.176570806" watchObservedRunningTime="2026-03-17 16:00:50.767939814 +0000 UTC m=+1442.181255861" Mar 17 16:00:51 crc kubenswrapper[4767]: I0317 16:00:51.798982 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"048fa918-95b2-404e-81f5-8989ed47ec0a","Type":"ContainerStarted","Data":"90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9"} Mar 17 16:00:51 crc kubenswrapper[4767]: I0317 16:00:51.804802 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e","Type":"ContainerStarted","Data":"db91577d9d962897e9869c843454246c300eb4ac87d6b82c10120049bf6d2916"} Mar 17 16:00:53 crc kubenswrapper[4767]: I0317 16:00:53.175766 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562720-bgk8g" Mar 17 16:00:53 crc kubenswrapper[4767]: I0317 16:00:53.333300 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p2sd\" (UniqueName: \"kubernetes.io/projected/93755875-2f62-4a18-aa46-a2c55496e1c2-kube-api-access-4p2sd\") pod \"93755875-2f62-4a18-aa46-a2c55496e1c2\" (UID: \"93755875-2f62-4a18-aa46-a2c55496e1c2\") " Mar 17 16:00:53 crc kubenswrapper[4767]: I0317 16:00:53.343110 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93755875-2f62-4a18-aa46-a2c55496e1c2-kube-api-access-4p2sd" (OuterVolumeSpecName: "kube-api-access-4p2sd") pod "93755875-2f62-4a18-aa46-a2c55496e1c2" (UID: "93755875-2f62-4a18-aa46-a2c55496e1c2"). InnerVolumeSpecName "kube-api-access-4p2sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:00:53 crc kubenswrapper[4767]: I0317 16:00:53.438913 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p2sd\" (UniqueName: \"kubernetes.io/projected/93755875-2f62-4a18-aa46-a2c55496e1c2-kube-api-access-4p2sd\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:53 crc kubenswrapper[4767]: I0317 16:00:53.842747 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"3ad3d357-572b-4b20-bfe6-4aa0b3513d41","Type":"ContainerStarted","Data":"bfb2e2fa71eea606d992e9bc0235bcb0151201ab641f392b4e91ac2749bdc185"} Mar 17 16:00:53 crc kubenswrapper[4767]: I0317 16:00:53.847867 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f","Type":"ContainerStarted","Data":"5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0"} Mar 17 16:00:53 crc kubenswrapper[4767]: I0317 16:00:53.854827 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562720-bgk8g" event={"ID":"93755875-2f62-4a18-aa46-a2c55496e1c2","Type":"ContainerDied","Data":"18c291a3f5a129afdd5ef20b479487cb0514533d490afccf556dabb91fdb12fb"} Mar 17 16:00:53 crc kubenswrapper[4767]: I0317 16:00:53.854873 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18c291a3f5a129afdd5ef20b479487cb0514533d490afccf556dabb91fdb12fb" Mar 17 16:00:53 crc kubenswrapper[4767]: I0317 16:00:53.854904 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562720-bgk8g" Mar 17 16:00:53 crc kubenswrapper[4767]: E0317 16:00:53.991312 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="8a122d32-81b0-48a8-a327-06be6bb37899" Mar 17 16:00:54 crc kubenswrapper[4767]: E0317 16:00:54.063825 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="4e7e6999-cfc1-4891-bbb4-48e848d2d8ef" Mar 17 16:00:54 crc kubenswrapper[4767]: I0317 16:00:54.268868 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562714-xlm92"] Mar 17 16:00:54 crc kubenswrapper[4767]: I0317 16:00:54.276602 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562714-xlm92"] Mar 17 16:00:54 crc kubenswrapper[4767]: I0317 16:00:54.866556 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8a122d32-81b0-48a8-a327-06be6bb37899","Type":"ContainerStarted","Data":"b3c5451bf2da33b563fb62272ab71ac501967d0a72ee38ca42f431ff84ffa6cf"} Mar 17 16:00:54 crc kubenswrapper[4767]: E0317 16:00:54.868189 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="8a122d32-81b0-48a8-a327-06be6bb37899" Mar 17 16:00:54 crc kubenswrapper[4767]: I0317 16:00:54.868819 4767 generic.go:334] "Generic (PLEG): container finished" podID="24aea215-34c0-40a4-a6fd-32a24222a50a" containerID="606579a80c63054628723eab95ac85853aa51d140a7eb024b13b0ed47d772f63" exitCode=0 Mar 17 16:00:54 crc kubenswrapper[4767]: I0317 16:00:54.868852 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" event={"ID":"24aea215-34c0-40a4-a6fd-32a24222a50a","Type":"ContainerDied","Data":"606579a80c63054628723eab95ac85853aa51d140a7eb024b13b0ed47d772f63"} Mar 17 16:00:54 crc kubenswrapper[4767]: I0317 16:00:54.872331 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271","Type":"ContainerStarted","Data":"6c8d3df4ac01b24cd441e147bfd08952987bed49b2332823d19d8d17d431abde"} Mar 17 16:00:54 crc kubenswrapper[4767]: I0317 16:00:54.881487 4767 generic.go:334] "Generic (PLEG): container finished" podID="b2781552-ff04-4f51-8901-7abb87dd1369" containerID="0bd89e5257dfbc11b4b35aafb16fd4e99dcbf69ba863634b70c351baf8ab7303" exitCode=0 Mar 17 16:00:54 crc kubenswrapper[4767]: I0317 16:00:54.881608 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" event={"ID":"b2781552-ff04-4f51-8901-7abb87dd1369","Type":"ContainerDied","Data":"0bd89e5257dfbc11b4b35aafb16fd4e99dcbf69ba863634b70c351baf8ab7303"} Mar 17 16:00:54 crc kubenswrapper[4767]: I0317 16:00:54.884818 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef","Type":"ContainerStarted","Data":"40e2a50355e284b6c9631faaacef63cd30316966d40c09f7a1a7f54ec9df23e3"} Mar 17 16:00:54 crc kubenswrapper[4767]: E0317 16:00:54.886427 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="4e7e6999-cfc1-4891-bbb4-48e848d2d8ef" Mar 17 16:00:54 crc kubenswrapper[4767]: I0317 16:00:54.909551 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-42gbs" event={"ID":"12012344-284e-4ce3-949e-dcd3812587ee","Type":"ContainerStarted","Data":"ebdb5712bfda8d2fa1e007da0174a08c862c650dec761572310ba699960d30f5"} Mar 17 16:00:54 crc kubenswrapper[4767]: I0317 16:00:54.919478 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6abe028c-416e-4978-bcc8-3a7b8d92624b","Type":"ContainerStarted","Data":"83147bdb48f8307a7017dec29264e33b9c5974cf82515fd5ce06e5234adb7e56"} Mar 17 16:00:55 crc kubenswrapper[4767]: I0317 16:00:55.001210 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-42gbs" podStartSLOduration=7.006085582 podStartE2EDuration="11.001163135s" podCreationTimestamp="2026-03-17 16:00:44 +0000 UTC" firstStartedPulling="2026-03-17 16:00:49.592160872 +0000 UTC m=+1441.005476919" lastFinishedPulling="2026-03-17 16:00:53.587238415 +0000 UTC m=+1445.000554472" observedRunningTime="2026-03-17 16:00:54.988934678 +0000 UTC m=+1446.402250725" watchObservedRunningTime="2026-03-17 16:00:55.001163135 +0000 UTC m=+1446.414479192" Mar 17 16:00:55 crc kubenswrapper[4767]: I0317 16:00:55.380549 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae556c0b-e4f1-42ce-98de-97e6241f8b5d" path="/var/lib/kubelet/pods/ae556c0b-e4f1-42ce-98de-97e6241f8b5d/volumes" Mar 17 16:00:55 crc kubenswrapper[4767]: I0317 16:00:55.950368 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a082b2bf-8b94-40e7-be0c-be64f75a4c3e","Type":"ContainerStarted","Data":"a150a9a169d4d790c0367da0e3743c173a7add3e0702539d16e0be602f1d139a"} Mar 17 16:00:55 crc kubenswrapper[4767]: I0317 16:00:55.955239 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" event={"ID":"24aea215-34c0-40a4-a6fd-32a24222a50a","Type":"ContainerStarted","Data":"43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3"} Mar 17 16:00:55 crc kubenswrapper[4767]: I0317 16:00:55.955452 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:00:55 crc kubenswrapper[4767]: I0317 16:00:55.957105 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kqr5p" event={"ID":"2aa82829-5daf-458e-b90d-705edcbbd3f2","Type":"ContainerStarted","Data":"10ffd31fcfad1f969c3a7f6bf44741f1b440a7942f0774db0cbdbcd1db548290"} Mar 17 16:00:55 crc kubenswrapper[4767]: I0317 16:00:55.959066 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" event={"ID":"b2781552-ff04-4f51-8901-7abb87dd1369","Type":"ContainerStarted","Data":"8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285"} Mar 17 16:00:55 crc kubenswrapper[4767]: E0317 16:00:55.961163 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="8a122d32-81b0-48a8-a327-06be6bb37899" Mar 17 16:00:55 crc kubenswrapper[4767]: E0317 16:00:55.961889 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="4e7e6999-cfc1-4891-bbb4-48e848d2d8ef" Mar 17 16:00:55 crc kubenswrapper[4767]: I0317 16:00:55.975083 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:00:56 crc kubenswrapper[4767]: I0317 16:00:56.024582 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" podStartSLOduration=7.676777216 podStartE2EDuration="11.024561907s" podCreationTimestamp="2026-03-17 16:00:45 +0000 UTC" firstStartedPulling="2026-03-17 16:00:50.257951608 +0000 UTC m=+1441.671267645" lastFinishedPulling="2026-03-17 16:00:53.605736289 +0000 UTC m=+1445.019052336" observedRunningTime="2026-03-17 16:00:56.023481398 +0000 UTC m=+1447.436797445" watchObservedRunningTime="2026-03-17 16:00:56.024561907 +0000 UTC m=+1447.437877954" Mar 17 16:00:56 crc kubenswrapper[4767]: I0317 16:00:56.056586 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" podStartSLOduration=7.5836918319999995 podStartE2EDuration="11.056559942s" podCreationTimestamp="2026-03-17 16:00:45 +0000 UTC" firstStartedPulling="2026-03-17 16:00:50.119161793 +0000 UTC m=+1441.532477850" lastFinishedPulling="2026-03-17 16:00:53.592029913 +0000 UTC m=+1445.005345960" observedRunningTime="2026-03-17 16:00:56.046275287 +0000 UTC m=+1447.459591354" watchObservedRunningTime="2026-03-17 16:00:56.056559942 +0000 UTC m=+1447.469875989" Mar 17 16:00:56 crc kubenswrapper[4767]: I0317 16:00:56.970138 4767 generic.go:334] "Generic (PLEG): container finished" podID="2aa82829-5daf-458e-b90d-705edcbbd3f2" containerID="10ffd31fcfad1f969c3a7f6bf44741f1b440a7942f0774db0cbdbcd1db548290" exitCode=0 Mar 17 16:00:56 crc kubenswrapper[4767]: I0317 16:00:56.970206 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kqr5p" event={"ID":"2aa82829-5daf-458e-b90d-705edcbbd3f2","Type":"ContainerDied","Data":"10ffd31fcfad1f969c3a7f6bf44741f1b440a7942f0774db0cbdbcd1db548290"} Mar 17 16:00:58 crc kubenswrapper[4767]: I0317 16:00:58.005902 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fa1b7ae1-40bc-474b-9602-de66005bea1c","Type":"ContainerStarted","Data":"9dd868db6d580f6256a83532d73e3b5d8ac83fd616c7393161d33868ee1f225f"} Mar 17 16:00:58 crc kubenswrapper[4767]: I0317 16:00:58.006767 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 17 16:00:58 crc kubenswrapper[4767]: I0317 16:00:58.013216 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kqr5p" event={"ID":"2aa82829-5daf-458e-b90d-705edcbbd3f2","Type":"ContainerStarted","Data":"6c2ab815d9830f829be59a7b70620e26edfadb9d91c2eeb36cf7b5ca5ab3c6bb"} Mar 17 16:00:58 crc kubenswrapper[4767]: I0317 16:00:58.013277 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kqr5p" event={"ID":"2aa82829-5daf-458e-b90d-705edcbbd3f2","Type":"ContainerStarted","Data":"58062a2c870bfc563f81eb3acdd785eb7aca16346aeedaacb35b73b7b62cfb0a"} Mar 17 16:00:58 crc kubenswrapper[4767]: I0317 16:00:58.013493 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:58 crc kubenswrapper[4767]: I0317 16:00:58.038829 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.679981535 podStartE2EDuration="1m6.038806125s" podCreationTimestamp="2026-03-17 15:59:52 +0000 UTC" firstStartedPulling="2026-03-17 15:59:54.699840967 +0000 UTC m=+1386.113157014" lastFinishedPulling="2026-03-17 16:00:57.058665557 +0000 UTC m=+1448.471981604" observedRunningTime="2026-03-17 16:00:58.030644947 +0000 UTC m=+1449.443961014" watchObservedRunningTime="2026-03-17 16:00:58.038806125 +0000 UTC m=+1449.452122162" Mar 17 16:00:58 crc kubenswrapper[4767]: I0317 16:00:58.061533 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-kqr5p" podStartSLOduration=9.034505125999999 podStartE2EDuration="58.061501831s" podCreationTimestamp="2026-03-17 16:00:00 +0000 UTC" firstStartedPulling="2026-03-17 16:00:05.965161029 +0000 UTC m=+1397.378477076" lastFinishedPulling="2026-03-17 16:00:54.992157724 +0000 UTC m=+1446.405473781" observedRunningTime="2026-03-17 16:00:58.056454656 +0000 UTC m=+1449.469770723" watchObservedRunningTime="2026-03-17 16:00:58.061501831 +0000 UTC m=+1449.474817878" Mar 17 16:00:58 crc kubenswrapper[4767]: I0317 16:00:58.674790 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-5c7b759f69-p6dvt" podUID="63b9ae01-2768-463b-8aab-569c360824a8" containerName="console" containerID="cri-o://2273e1025467dcb4d2442bd5702080c625fd5a89a6060769921396296113365e" gracePeriod=15 Mar 17 16:00:58 crc kubenswrapper[4767]: E0317 16:00:58.913296 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63b9ae01_2768_463b_8aab_569c360824a8.slice/crio-conmon-2273e1025467dcb4d2442bd5702080c625fd5a89a6060769921396296113365e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63b9ae01_2768_463b_8aab_569c360824a8.slice/crio-2273e1025467dcb4d2442bd5702080c625fd5a89a6060769921396296113365e.scope\": RecentStats: unable to find data in memory cache]" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.033057 4767 generic.go:334] "Generic (PLEG): container finished" podID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" containerID="6c8d3df4ac01b24cd441e147bfd08952987bed49b2332823d19d8d17d431abde" exitCode=0 Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.033259 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271","Type":"ContainerDied","Data":"6c8d3df4ac01b24cd441e147bfd08952987bed49b2332823d19d8d17d431abde"} Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.056304 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5c7b759f69-p6dvt_63b9ae01-2768-463b-8aab-569c360824a8/console/0.log" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.056362 4767 generic.go:334] "Generic (PLEG): container finished" podID="63b9ae01-2768-463b-8aab-569c360824a8" containerID="2273e1025467dcb4d2442bd5702080c625fd5a89a6060769921396296113365e" exitCode=2 Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.056476 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c7b759f69-p6dvt" event={"ID":"63b9ae01-2768-463b-8aab-569c360824a8","Type":"ContainerDied","Data":"2273e1025467dcb4d2442bd5702080c625fd5a89a6060769921396296113365e"} Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.063270 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-47tcm" event={"ID":"d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0","Type":"ContainerStarted","Data":"8ccf1c1fb70053b81df80f05174976116414f59b6fee2deba39f010c85c62912"} Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.065516 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-47tcm" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.243810 4767 generic.go:334] "Generic (PLEG): container finished" podID="6abe028c-416e-4978-bcc8-3a7b8d92624b" containerID="83147bdb48f8307a7017dec29264e33b9c5974cf82515fd5ce06e5234adb7e56" exitCode=0 Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.245746 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6abe028c-416e-4978-bcc8-3a7b8d92624b","Type":"ContainerDied","Data":"83147bdb48f8307a7017dec29264e33b9c5974cf82515fd5ce06e5234adb7e56"} Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.246529 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.281114 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-47tcm" podStartSLOduration=5.8692711509999995 podStartE2EDuration="59.281067231s" podCreationTimestamp="2026-03-17 16:00:00 +0000 UTC" firstStartedPulling="2026-03-17 16:00:04.484291505 +0000 UTC m=+1395.897607542" lastFinishedPulling="2026-03-17 16:00:57.896087575 +0000 UTC m=+1449.309403622" observedRunningTime="2026-03-17 16:00:59.264773296 +0000 UTC m=+1450.678089363" watchObservedRunningTime="2026-03-17 16:00:59.281067231 +0000 UTC m=+1450.694383278" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.463808 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5c7b759f69-p6dvt_63b9ae01-2768-463b-8aab-569c360824a8/console/0.log" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.464049 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.634053 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-oauth-serving-cert\") pod \"63b9ae01-2768-463b-8aab-569c360824a8\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.634578 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-oauth-config\") pod \"63b9ae01-2768-463b-8aab-569c360824a8\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.634641 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-service-ca\") pod \"63b9ae01-2768-463b-8aab-569c360824a8\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.634683 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-trusted-ca-bundle\") pod \"63b9ae01-2768-463b-8aab-569c360824a8\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.634806 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-console-config\") pod \"63b9ae01-2768-463b-8aab-569c360824a8\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.635105 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "63b9ae01-2768-463b-8aab-569c360824a8" (UID: "63b9ae01-2768-463b-8aab-569c360824a8"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.635455 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-service-ca" (OuterVolumeSpecName: "service-ca") pod "63b9ae01-2768-463b-8aab-569c360824a8" (UID: "63b9ae01-2768-463b-8aab-569c360824a8"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.635523 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-console-config" (OuterVolumeSpecName: "console-config") pod "63b9ae01-2768-463b-8aab-569c360824a8" (UID: "63b9ae01-2768-463b-8aab-569c360824a8"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.635517 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "63b9ae01-2768-463b-8aab-569c360824a8" (UID: "63b9ae01-2768-463b-8aab-569c360824a8"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.635828 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzln9\" (UniqueName: \"kubernetes.io/projected/63b9ae01-2768-463b-8aab-569c360824a8-kube-api-access-dzln9\") pod \"63b9ae01-2768-463b-8aab-569c360824a8\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.635867 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-serving-cert\") pod \"63b9ae01-2768-463b-8aab-569c360824a8\" (UID: \"63b9ae01-2768-463b-8aab-569c360824a8\") " Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.636507 4767 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.636533 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.636542 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.636550 4767 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/63b9ae01-2768-463b-8aab-569c360824a8-console-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.641059 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63b9ae01-2768-463b-8aab-569c360824a8-kube-api-access-dzln9" (OuterVolumeSpecName: "kube-api-access-dzln9") pod "63b9ae01-2768-463b-8aab-569c360824a8" (UID: "63b9ae01-2768-463b-8aab-569c360824a8"). InnerVolumeSpecName "kube-api-access-dzln9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.642214 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "63b9ae01-2768-463b-8aab-569c360824a8" (UID: "63b9ae01-2768-463b-8aab-569c360824a8"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.642404 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "63b9ae01-2768-463b-8aab-569c360824a8" (UID: "63b9ae01-2768-463b-8aab-569c360824a8"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.739189 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzln9\" (UniqueName: \"kubernetes.io/projected/63b9ae01-2768-463b-8aab-569c360824a8-kube-api-access-dzln9\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.739246 4767 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 16:00:59 crc kubenswrapper[4767]: I0317 16:00:59.739264 4767 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/63b9ae01-2768-463b-8aab-569c360824a8-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.257449 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5c7b759f69-p6dvt_63b9ae01-2768-463b-8aab-569c360824a8/console/0.log" Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.257540 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c7b759f69-p6dvt" event={"ID":"63b9ae01-2768-463b-8aab-569c360824a8","Type":"ContainerDied","Data":"a003cd890c236aedefd25287e0942d73fe5daab8653ca1eaf738a15887ce6c7a"} Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.257584 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c7b759f69-p6dvt" Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.257596 4767 scope.go:117] "RemoveContainer" containerID="2273e1025467dcb4d2442bd5702080c625fd5a89a6060769921396296113365e" Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.262219 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6abe028c-416e-4978-bcc8-3a7b8d92624b","Type":"ContainerStarted","Data":"6abac3a757369ed7aa38da0ae38fc79753f588411284852e6d063714ed21a5fd"} Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.265236 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271","Type":"ContainerStarted","Data":"3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac"} Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.267069 4767 generic.go:334] "Generic (PLEG): container finished" podID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerID="5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0" exitCode=0 Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.267180 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f","Type":"ContainerDied","Data":"5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0"} Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.300327 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=11.584795658 podStartE2EDuration="1m8.300303233s" podCreationTimestamp="2026-03-17 15:59:52 +0000 UTC" firstStartedPulling="2026-03-17 15:59:56.89008379 +0000 UTC m=+1388.303399837" lastFinishedPulling="2026-03-17 16:00:53.605591365 +0000 UTC m=+1445.018907412" observedRunningTime="2026-03-17 16:01:00.289491314 +0000 UTC m=+1451.702807361" watchObservedRunningTime="2026-03-17 16:01:00.300303233 +0000 UTC m=+1451.713619270" Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.323222 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5c7b759f69-p6dvt"] Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.331064 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-5c7b759f69-p6dvt"] Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.367883 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=10.848854073 podStartE2EDuration="1m10.367859947s" podCreationTimestamp="2026-03-17 15:59:50 +0000 UTC" firstStartedPulling="2026-03-17 15:59:54.088372399 +0000 UTC m=+1385.501688446" lastFinishedPulling="2026-03-17 16:00:53.607378273 +0000 UTC m=+1445.020694320" observedRunningTime="2026-03-17 16:01:00.35861612 +0000 UTC m=+1451.771932177" watchObservedRunningTime="2026-03-17 16:01:00.367859947 +0000 UTC m=+1451.781175994" Mar 17 16:01:00 crc kubenswrapper[4767]: I0317 16:01:00.980438 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:01:01 crc kubenswrapper[4767]: I0317 16:01:01.024974 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:01:01 crc kubenswrapper[4767]: I0317 16:01:01.061063 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xhq4t"] Mar 17 16:01:01 crc kubenswrapper[4767]: I0317 16:01:01.279334 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" podUID="24aea215-34c0-40a4-a6fd-32a24222a50a" containerName="dnsmasq-dns" containerID="cri-o://43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3" gracePeriod=10 Mar 17 16:01:01 crc kubenswrapper[4767]: I0317 16:01:01.368448 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63b9ae01-2768-463b-8aab-569c360824a8" path="/var/lib/kubelet/pods/63b9ae01-2768-463b-8aab-569c360824a8/volumes" Mar 17 16:01:01 crc kubenswrapper[4767]: I0317 16:01:01.915505 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:01:01 crc kubenswrapper[4767]: I0317 16:01:01.996199 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-config\") pod \"24aea215-34c0-40a4-a6fd-32a24222a50a\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " Mar 17 16:01:01 crc kubenswrapper[4767]: I0317 16:01:01.996298 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-dns-svc\") pod \"24aea215-34c0-40a4-a6fd-32a24222a50a\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " Mar 17 16:01:01 crc kubenswrapper[4767]: I0317 16:01:01.996411 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-ovsdbserver-nb\") pod \"24aea215-34c0-40a4-a6fd-32a24222a50a\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " Mar 17 16:01:01 crc kubenswrapper[4767]: I0317 16:01:01.996469 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcr5w\" (UniqueName: \"kubernetes.io/projected/24aea215-34c0-40a4-a6fd-32a24222a50a-kube-api-access-dcr5w\") pod \"24aea215-34c0-40a4-a6fd-32a24222a50a\" (UID: \"24aea215-34c0-40a4-a6fd-32a24222a50a\") " Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.013215 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24aea215-34c0-40a4-a6fd-32a24222a50a-kube-api-access-dcr5w" (OuterVolumeSpecName: "kube-api-access-dcr5w") pod "24aea215-34c0-40a4-a6fd-32a24222a50a" (UID: "24aea215-34c0-40a4-a6fd-32a24222a50a"). InnerVolumeSpecName "kube-api-access-dcr5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.100242 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcr5w\" (UniqueName: \"kubernetes.io/projected/24aea215-34c0-40a4-a6fd-32a24222a50a-kube-api-access-dcr5w\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.116933 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "24aea215-34c0-40a4-a6fd-32a24222a50a" (UID: "24aea215-34c0-40a4-a6fd-32a24222a50a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.128045 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "24aea215-34c0-40a4-a6fd-32a24222a50a" (UID: "24aea215-34c0-40a4-a6fd-32a24222a50a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.130984 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-config" (OuterVolumeSpecName: "config") pod "24aea215-34c0-40a4-a6fd-32a24222a50a" (UID: "24aea215-34c0-40a4-a6fd-32a24222a50a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.203988 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.204397 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.204464 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24aea215-34c0-40a4-a6fd-32a24222a50a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.295229 4767 generic.go:334] "Generic (PLEG): container finished" podID="24aea215-34c0-40a4-a6fd-32a24222a50a" containerID="43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3" exitCode=0 Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.295317 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" event={"ID":"24aea215-34c0-40a4-a6fd-32a24222a50a","Type":"ContainerDied","Data":"43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3"} Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.295352 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" event={"ID":"24aea215-34c0-40a4-a6fd-32a24222a50a","Type":"ContainerDied","Data":"cdbc9bea367092e702d46ead0f629d6c4fbf64a65f246ac40495d6826ce18e70"} Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.295407 4767 scope.go:117] "RemoveContainer" containerID="43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.295669 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-xhq4t" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.330446 4767 scope.go:117] "RemoveContainer" containerID="606579a80c63054628723eab95ac85853aa51d140a7eb024b13b0ed47d772f63" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.342071 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xhq4t"] Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.351528 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xhq4t"] Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.359089 4767 scope.go:117] "RemoveContainer" containerID="43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3" Mar 17 16:01:02 crc kubenswrapper[4767]: E0317 16:01:02.359842 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3\": container with ID starting with 43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3 not found: ID does not exist" containerID="43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.359892 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3"} err="failed to get container status \"43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3\": rpc error: code = NotFound desc = could not find container \"43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3\": container with ID starting with 43f0a063132a5460f5e581608b50ec90b4ad6abfdd308f11e9f2c81c13e249d3 not found: ID does not exist" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.359922 4767 scope.go:117] "RemoveContainer" containerID="606579a80c63054628723eab95ac85853aa51d140a7eb024b13b0ed47d772f63" Mar 17 16:01:02 crc kubenswrapper[4767]: E0317 16:01:02.361764 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"606579a80c63054628723eab95ac85853aa51d140a7eb024b13b0ed47d772f63\": container with ID starting with 606579a80c63054628723eab95ac85853aa51d140a7eb024b13b0ed47d772f63 not found: ID does not exist" containerID="606579a80c63054628723eab95ac85853aa51d140a7eb024b13b0ed47d772f63" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.361798 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"606579a80c63054628723eab95ac85853aa51d140a7eb024b13b0ed47d772f63"} err="failed to get container status \"606579a80c63054628723eab95ac85853aa51d140a7eb024b13b0ed47d772f63\": rpc error: code = NotFound desc = could not find container \"606579a80c63054628723eab95ac85853aa51d140a7eb024b13b0ed47d772f63\": container with ID starting with 606579a80c63054628723eab95ac85853aa51d140a7eb024b13b0ed47d772f63 not found: ID does not exist" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.489839 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 17 16:01:02 crc kubenswrapper[4767]: I0317 16:01:02.489899 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 17 16:01:02 crc kubenswrapper[4767]: E0317 16:01:02.573221 4767 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.119:59952->38.102.83.119:38263: write tcp 38.102.83.119:59952->38.102.83.119:38263: write: broken pipe Mar 17 16:01:03 crc kubenswrapper[4767]: I0317 16:01:03.310847 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"232c4872-a09a-4be2-a7d2-c3c7232bae8a","Type":"ContainerStarted","Data":"74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e"} Mar 17 16:01:03 crc kubenswrapper[4767]: I0317 16:01:03.311675 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 17 16:01:03 crc kubenswrapper[4767]: I0317 16:01:03.317528 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 17 16:01:03 crc kubenswrapper[4767]: I0317 16:01:03.338929 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=5.860013052 podStartE2EDuration="1m5.338896879s" podCreationTimestamp="2026-03-17 15:59:58 +0000 UTC" firstStartedPulling="2026-03-17 16:00:03.283841779 +0000 UTC m=+1394.697157826" lastFinishedPulling="2026-03-17 16:01:02.762725616 +0000 UTC m=+1454.176041653" observedRunningTime="2026-03-17 16:01:03.33219332 +0000 UTC m=+1454.745509377" watchObservedRunningTime="2026-03-17 16:01:03.338896879 +0000 UTC m=+1454.752212926" Mar 17 16:01:03 crc kubenswrapper[4767]: I0317 16:01:03.376580 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24aea215-34c0-40a4-a6fd-32a24222a50a" path="/var/lib/kubelet/pods/24aea215-34c0-40a4-a6fd-32a24222a50a/volumes" Mar 17 16:01:03 crc kubenswrapper[4767]: I0317 16:01:03.929773 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 17 16:01:03 crc kubenswrapper[4767]: I0317 16:01:03.930126 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 17 16:01:04 crc kubenswrapper[4767]: I0317 16:01:04.367454 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 17 16:01:04 crc kubenswrapper[4767]: I0317 16:01:04.464892 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 17 16:01:06 crc kubenswrapper[4767]: I0317 16:01:06.689102 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 17 16:01:06 crc kubenswrapper[4767]: I0317 16:01:06.775869 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.292541 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-tt884"] Mar 17 16:01:07 crc kubenswrapper[4767]: E0317 16:01:07.293127 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b9ae01-2768-463b-8aab-569c360824a8" containerName="console" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.293152 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b9ae01-2768-463b-8aab-569c360824a8" containerName="console" Mar 17 16:01:07 crc kubenswrapper[4767]: E0317 16:01:07.293214 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93755875-2f62-4a18-aa46-a2c55496e1c2" containerName="oc" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.293223 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93755875-2f62-4a18-aa46-a2c55496e1c2" containerName="oc" Mar 17 16:01:07 crc kubenswrapper[4767]: E0317 16:01:07.293241 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24aea215-34c0-40a4-a6fd-32a24222a50a" containerName="init" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.293248 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="24aea215-34c0-40a4-a6fd-32a24222a50a" containerName="init" Mar 17 16:01:07 crc kubenswrapper[4767]: E0317 16:01:07.293266 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24aea215-34c0-40a4-a6fd-32a24222a50a" containerName="dnsmasq-dns" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.293273 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="24aea215-34c0-40a4-a6fd-32a24222a50a" containerName="dnsmasq-dns" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.293563 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="24aea215-34c0-40a4-a6fd-32a24222a50a" containerName="dnsmasq-dns" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.293595 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93755875-2f62-4a18-aa46-a2c55496e1c2" containerName="oc" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.293611 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="63b9ae01-2768-463b-8aab-569c360824a8" containerName="console" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.295295 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.318220 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tt884"] Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.367963 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.368394 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-config\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.369600 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbpr6\" (UniqueName: \"kubernetes.io/projected/788da75a-4303-4878-a388-217e6beee11b-kube-api-access-jbpr6\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.370382 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-dns-svc\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.370425 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.474469 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.474533 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-config\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.474848 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbpr6\" (UniqueName: \"kubernetes.io/projected/788da75a-4303-4878-a388-217e6beee11b-kube-api-access-jbpr6\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.475153 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-dns-svc\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.475322 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.475861 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-config\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.475966 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.476827 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.476898 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-dns-svc\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.504965 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbpr6\" (UniqueName: \"kubernetes.io/projected/788da75a-4303-4878-a388-217e6beee11b-kube-api-access-jbpr6\") pod \"dnsmasq-dns-698758b865-tt884\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:07 crc kubenswrapper[4767]: I0317 16:01:07.631266 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.445387 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.533721 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.534100 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.539131 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.539652 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.540243 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-wtjtt" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.541018 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.734900 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/803604b0-bf06-4c1b-bdce-338284abc323-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.734965 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/803604b0-bf06-4c1b-bdce-338284abc323-lock\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.735180 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/803604b0-bf06-4c1b-bdce-338284abc323-cache\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.735259 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.735294 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh2xw\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-kube-api-access-gh2xw\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.735327 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e9bd84aa-620b-483d-9c78-e9394d39bf23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9bd84aa-620b-483d-9c78-e9394d39bf23\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.837603 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/803604b0-bf06-4c1b-bdce-338284abc323-cache\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.837720 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.837764 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh2xw\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-kube-api-access-gh2xw\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.837802 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e9bd84aa-620b-483d-9c78-e9394d39bf23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9bd84aa-620b-483d-9c78-e9394d39bf23\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: E0317 16:01:08.837925 4767 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 16:01:08 crc kubenswrapper[4767]: E0317 16:01:08.837952 4767 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 16:01:08 crc kubenswrapper[4767]: E0317 16:01:08.838009 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift podName:803604b0-bf06-4c1b-bdce-338284abc323 nodeName:}" failed. No retries permitted until 2026-03-17 16:01:09.337991187 +0000 UTC m=+1460.751307234 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift") pod "swift-storage-0" (UID: "803604b0-bf06-4c1b-bdce-338284abc323") : configmap "swift-ring-files" not found Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.837932 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/803604b0-bf06-4c1b-bdce-338284abc323-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.838200 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/803604b0-bf06-4c1b-bdce-338284abc323-lock\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.838857 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/803604b0-bf06-4c1b-bdce-338284abc323-cache\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.838932 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/803604b0-bf06-4c1b-bdce-338284abc323-lock\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.842604 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.842644 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e9bd84aa-620b-483d-9c78-e9394d39bf23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9bd84aa-620b-483d-9c78-e9394d39bf23\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d2089ddf478fcaf7a0c35ea4502dec8a62f9f3bdfbf0482d6eefa97561d2d582/globalmount\"" pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.853339 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/803604b0-bf06-4c1b-bdce-338284abc323-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.869424 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh2xw\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-kube-api-access-gh2xw\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:08 crc kubenswrapper[4767]: I0317 16:01:08.894498 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e9bd84aa-620b-483d-9c78-e9394d39bf23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9bd84aa-620b-483d-9c78-e9394d39bf23\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.046807 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vhvs8"] Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.048393 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.074255 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vhvs8"] Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.118283 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-cqcdm"] Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.125363 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.128725 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.128862 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.130630 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.135125 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-cqcdm"] Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.248829 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-combined-ca-bundle\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.248921 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-dispersionconf\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.248977 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6znwv\" (UniqueName: \"kubernetes.io/projected/500b62ad-a562-419e-a44a-85030e12bf8b-kube-api-access-6znwv\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.249197 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-swiftconf\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.249298 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-scripts\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.249612 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-ring-data-devices\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.249672 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82b61927-c834-4a58-a140-c2c63686f6f1-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-vhvs8\" (UID: \"82b61927-c834-4a58-a140-c2c63686f6f1\") " pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.249729 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/500b62ad-a562-419e-a44a-85030e12bf8b-etc-swift\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.251082 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cl7c\" (UniqueName: \"kubernetes.io/projected/82b61927-c834-4a58-a140-c2c63686f6f1-kube-api-access-6cl7c\") pod \"mysqld-exporter-openstack-db-create-vhvs8\" (UID: \"82b61927-c834-4a58-a140-c2c63686f6f1\") " pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.354374 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cl7c\" (UniqueName: \"kubernetes.io/projected/82b61927-c834-4a58-a140-c2c63686f6f1-kube-api-access-6cl7c\") pod \"mysqld-exporter-openstack-db-create-vhvs8\" (UID: \"82b61927-c834-4a58-a140-c2c63686f6f1\") " pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.354584 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-combined-ca-bundle\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.354664 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-dispersionconf\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.354737 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6znwv\" (UniqueName: \"kubernetes.io/projected/500b62ad-a562-419e-a44a-85030e12bf8b-kube-api-access-6znwv\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.354837 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.354903 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-swiftconf\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.354966 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-scripts\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: E0317 16:01:09.354995 4767 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 16:01:09 crc kubenswrapper[4767]: E0317 16:01:09.355018 4767 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 16:01:09 crc kubenswrapper[4767]: E0317 16:01:09.355075 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift podName:803604b0-bf06-4c1b-bdce-338284abc323 nodeName:}" failed. No retries permitted until 2026-03-17 16:01:10.355054162 +0000 UTC m=+1461.768370209 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift") pod "swift-storage-0" (UID: "803604b0-bf06-4c1b-bdce-338284abc323") : configmap "swift-ring-files" not found Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.355114 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-ring-data-devices\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.355157 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82b61927-c834-4a58-a140-c2c63686f6f1-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-vhvs8\" (UID: \"82b61927-c834-4a58-a140-c2c63686f6f1\") " pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.355227 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/500b62ad-a562-419e-a44a-85030e12bf8b-etc-swift\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.355986 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-scripts\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.356445 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-ring-data-devices\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.356736 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82b61927-c834-4a58-a140-c2c63686f6f1-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-vhvs8\" (UID: \"82b61927-c834-4a58-a140-c2c63686f6f1\") " pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.356848 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/500b62ad-a562-419e-a44a-85030e12bf8b-etc-swift\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.365139 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-swiftconf\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.373228 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-dispersionconf\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.373645 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-combined-ca-bundle\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.374998 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cl7c\" (UniqueName: \"kubernetes.io/projected/82b61927-c834-4a58-a140-c2c63686f6f1-kube-api-access-6cl7c\") pod \"mysqld-exporter-openstack-db-create-vhvs8\" (UID: \"82b61927-c834-4a58-a140-c2c63686f6f1\") " pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.376114 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6znwv\" (UniqueName: \"kubernetes.io/projected/500b62ad-a562-419e-a44a-85030e12bf8b-kube-api-access-6znwv\") pod \"swift-ring-rebalance-cqcdm\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.450810 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.478058 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 17 16:01:09 crc kubenswrapper[4767]: I0317 16:01:09.670313 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" Mar 17 16:01:10 crc kubenswrapper[4767]: I0317 16:01:10.214596 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tt884"] Mar 17 16:01:10 crc kubenswrapper[4767]: I0317 16:01:10.364851 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vhvs8"] Mar 17 16:01:10 crc kubenswrapper[4767]: I0317 16:01:10.398931 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:10 crc kubenswrapper[4767]: E0317 16:01:10.400403 4767 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 16:01:10 crc kubenswrapper[4767]: E0317 16:01:10.400429 4767 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 16:01:10 crc kubenswrapper[4767]: E0317 16:01:10.400498 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift podName:803604b0-bf06-4c1b-bdce-338284abc323 nodeName:}" failed. No retries permitted until 2026-03-17 16:01:12.400477922 +0000 UTC m=+1463.813793969 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift") pod "swift-storage-0" (UID: "803604b0-bf06-4c1b-bdce-338284abc323") : configmap "swift-ring-files" not found Mar 17 16:01:10 crc kubenswrapper[4767]: I0317 16:01:10.423043 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-cqcdm"] Mar 17 16:01:10 crc kubenswrapper[4767]: I0317 16:01:10.482227 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tt884" event={"ID":"788da75a-4303-4878-a388-217e6beee11b","Type":"ContainerStarted","Data":"aa3eb1ef2027a57c8a0a5526e34a64782c7b406e6bb553cfd3327ad757a8f18a"} Mar 17 16:01:10 crc kubenswrapper[4767]: I0317 16:01:10.484798 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-cqcdm" event={"ID":"500b62ad-a562-419e-a44a-85030e12bf8b","Type":"ContainerStarted","Data":"db38d47c5fda62600a620d5a2a323a5cd2e83faf6ff7061d977148ec4cbfb4a2"} Mar 17 16:01:10 crc kubenswrapper[4767]: I0317 16:01:10.488523 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" event={"ID":"82b61927-c834-4a58-a140-c2c63686f6f1","Type":"ContainerStarted","Data":"465f6192f422eac5b7e0151f29891ab6b8ad6b9629d06c1080087fd1eac41f50"} Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.502475 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8a122d32-81b0-48a8-a327-06be6bb37899","Type":"ContainerStarted","Data":"db1b27823150871307e9db1a2b3d79c3b52ebbb6bb1c46e3a082ef8969e17b7f"} Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.508543 4767 generic.go:334] "Generic (PLEG): container finished" podID="82b61927-c834-4a58-a140-c2c63686f6f1" containerID="0bd8254376f5c2166be6220136c28289e2241312119ecc0a6eae2580e67c364d" exitCode=0 Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.508617 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" event={"ID":"82b61927-c834-4a58-a140-c2c63686f6f1","Type":"ContainerDied","Data":"0bd8254376f5c2166be6220136c28289e2241312119ecc0a6eae2580e67c364d"} Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.511773 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4e7e6999-cfc1-4891-bbb4-48e848d2d8ef","Type":"ContainerStarted","Data":"c8e74a9babd832702d3bebe05edce1f8f97988d7254d33971802a77f9c610f7a"} Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.516015 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f","Type":"ContainerStarted","Data":"9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616"} Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.520392 4767 generic.go:334] "Generic (PLEG): container finished" podID="788da75a-4303-4878-a388-217e6beee11b" containerID="9ede4241b1f554851e9b2d432b9fc848e9d31ef9c6c6dff53102be9491b7f53b" exitCode=0 Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.520445 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tt884" event={"ID":"788da75a-4303-4878-a388-217e6beee11b","Type":"ContainerDied","Data":"9ede4241b1f554851e9b2d432b9fc848e9d31ef9c6c6dff53102be9491b7f53b"} Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.558966 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=12.495141638 podStartE2EDuration="1m8.558945382s" podCreationTimestamp="2026-03-17 16:00:03 +0000 UTC" firstStartedPulling="2026-03-17 16:00:13.701712955 +0000 UTC m=+1405.115029002" lastFinishedPulling="2026-03-17 16:01:09.765516699 +0000 UTC m=+1461.178832746" observedRunningTime="2026-03-17 16:01:11.531688184 +0000 UTC m=+1462.945004241" watchObservedRunningTime="2026-03-17 16:01:11.558945382 +0000 UTC m=+1462.972261429" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.627609 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.268029131 podStartE2EDuration="1m8.627583554s" podCreationTimestamp="2026-03-17 16:00:03 +0000 UTC" firstStartedPulling="2026-03-17 16:00:13.693346578 +0000 UTC m=+1405.106662625" lastFinishedPulling="2026-03-17 16:01:11.052901001 +0000 UTC m=+1462.466217048" observedRunningTime="2026-03-17 16:01:11.622961871 +0000 UTC m=+1463.036277928" watchObservedRunningTime="2026-03-17 16:01:11.627583554 +0000 UTC m=+1463.040899601" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.693545 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-ccf0-account-create-update-vzhjl"] Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.696690 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.703017 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.711538 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-ccf0-account-create-update-vzhjl"] Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.728990 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-z5gph"] Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.731699 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-z5gph" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.735448 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.770117 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.770285 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.796897 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-z5gph"] Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.880677 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2s6g\" (UniqueName: \"kubernetes.io/projected/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-kube-api-access-l2s6g\") pod \"root-account-create-update-z5gph\" (UID: \"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c\") " pod="openstack/root-account-create-update-z5gph" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.881187 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6gm5\" (UniqueName: \"kubernetes.io/projected/22e12d8e-e55f-4839-afdb-a03c60523aed-kube-api-access-l6gm5\") pod \"mysqld-exporter-ccf0-account-create-update-vzhjl\" (UID: \"22e12d8e-e55f-4839-afdb-a03c60523aed\") " pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.881287 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-operator-scripts\") pod \"root-account-create-update-z5gph\" (UID: \"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c\") " pod="openstack/root-account-create-update-z5gph" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.881465 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22e12d8e-e55f-4839-afdb-a03c60523aed-operator-scripts\") pod \"mysqld-exporter-ccf0-account-create-update-vzhjl\" (UID: \"22e12d8e-e55f-4839-afdb-a03c60523aed\") " pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.984109 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6gm5\" (UniqueName: \"kubernetes.io/projected/22e12d8e-e55f-4839-afdb-a03c60523aed-kube-api-access-l6gm5\") pod \"mysqld-exporter-ccf0-account-create-update-vzhjl\" (UID: \"22e12d8e-e55f-4839-afdb-a03c60523aed\") " pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.984270 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-operator-scripts\") pod \"root-account-create-update-z5gph\" (UID: \"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c\") " pod="openstack/root-account-create-update-z5gph" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.984353 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22e12d8e-e55f-4839-afdb-a03c60523aed-operator-scripts\") pod \"mysqld-exporter-ccf0-account-create-update-vzhjl\" (UID: \"22e12d8e-e55f-4839-afdb-a03c60523aed\") " pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.984450 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2s6g\" (UniqueName: \"kubernetes.io/projected/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-kube-api-access-l2s6g\") pod \"root-account-create-update-z5gph\" (UID: \"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c\") " pod="openstack/root-account-create-update-z5gph" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.986099 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-operator-scripts\") pod \"root-account-create-update-z5gph\" (UID: \"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c\") " pod="openstack/root-account-create-update-z5gph" Mar 17 16:01:11 crc kubenswrapper[4767]: I0317 16:01:11.986868 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22e12d8e-e55f-4839-afdb-a03c60523aed-operator-scripts\") pod \"mysqld-exporter-ccf0-account-create-update-vzhjl\" (UID: \"22e12d8e-e55f-4839-afdb-a03c60523aed\") " pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" Mar 17 16:01:12 crc kubenswrapper[4767]: I0317 16:01:12.006004 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2s6g\" (UniqueName: \"kubernetes.io/projected/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-kube-api-access-l2s6g\") pod \"root-account-create-update-z5gph\" (UID: \"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c\") " pod="openstack/root-account-create-update-z5gph" Mar 17 16:01:12 crc kubenswrapper[4767]: I0317 16:01:12.006581 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6gm5\" (UniqueName: \"kubernetes.io/projected/22e12d8e-e55f-4839-afdb-a03c60523aed-kube-api-access-l6gm5\") pod \"mysqld-exporter-ccf0-account-create-update-vzhjl\" (UID: \"22e12d8e-e55f-4839-afdb-a03c60523aed\") " pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" Mar 17 16:01:12 crc kubenswrapper[4767]: I0317 16:01:12.027739 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" Mar 17 16:01:12 crc kubenswrapper[4767]: I0317 16:01:12.067123 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-z5gph" Mar 17 16:01:12 crc kubenswrapper[4767]: I0317 16:01:12.412535 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:12 crc kubenswrapper[4767]: E0317 16:01:12.412773 4767 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 16:01:12 crc kubenswrapper[4767]: E0317 16:01:12.413293 4767 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 16:01:12 crc kubenswrapper[4767]: E0317 16:01:12.413369 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift podName:803604b0-bf06-4c1b-bdce-338284abc323 nodeName:}" failed. No retries permitted until 2026-03-17 16:01:16.413349423 +0000 UTC m=+1467.826665470 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift") pod "swift-storage-0" (UID: "803604b0-bf06-4c1b-bdce-338284abc323") : configmap "swift-ring-files" not found Mar 17 16:01:12 crc kubenswrapper[4767]: I0317 16:01:12.540522 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tt884" event={"ID":"788da75a-4303-4878-a388-217e6beee11b","Type":"ContainerStarted","Data":"f926d0748efcabdaa65f2e4500ba9d42c0ab10d5476dd9dd4e281cd161bfe90b"} Mar 17 16:01:12 crc kubenswrapper[4767]: I0317 16:01:12.574949 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-tt884" podStartSLOduration=5.574918857 podStartE2EDuration="5.574918857s" podCreationTimestamp="2026-03-17 16:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:01:12.56603912 +0000 UTC m=+1463.979355197" watchObservedRunningTime="2026-03-17 16:01:12.574918857 +0000 UTC m=+1463.988234914" Mar 17 16:01:12 crc kubenswrapper[4767]: I0317 16:01:12.596100 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-ccf0-account-create-update-vzhjl"] Mar 17 16:01:12 crc kubenswrapper[4767]: I0317 16:01:12.632291 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:12 crc kubenswrapper[4767]: I0317 16:01:12.725463 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-z5gph"] Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.536783 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-kplsg"] Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.538837 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kplsg" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.546324 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvv6q\" (UniqueName: \"kubernetes.io/projected/33ed389a-fa13-445f-826e-a88c04ef059d-kube-api-access-fvv6q\") pod \"glance-db-create-kplsg\" (UID: \"33ed389a-fa13-445f-826e-a88c04ef059d\") " pod="openstack/glance-db-create-kplsg" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.546524 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33ed389a-fa13-445f-826e-a88c04ef059d-operator-scripts\") pod \"glance-db-create-kplsg\" (UID: \"33ed389a-fa13-445f-826e-a88c04ef059d\") " pod="openstack/glance-db-create-kplsg" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.560601 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kplsg"] Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.646134 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-c428-account-create-update-x48t4"] Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.647886 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33ed389a-fa13-445f-826e-a88c04ef059d-operator-scripts\") pod \"glance-db-create-kplsg\" (UID: \"33ed389a-fa13-445f-826e-a88c04ef059d\") " pod="openstack/glance-db-create-kplsg" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.648084 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvv6q\" (UniqueName: \"kubernetes.io/projected/33ed389a-fa13-445f-826e-a88c04ef059d-kube-api-access-fvv6q\") pod \"glance-db-create-kplsg\" (UID: \"33ed389a-fa13-445f-826e-a88c04ef059d\") " pod="openstack/glance-db-create-kplsg" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.648235 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c428-account-create-update-x48t4" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.649339 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33ed389a-fa13-445f-826e-a88c04ef059d-operator-scripts\") pod \"glance-db-create-kplsg\" (UID: \"33ed389a-fa13-445f-826e-a88c04ef059d\") " pod="openstack/glance-db-create-kplsg" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.652060 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.655329 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c428-account-create-update-x48t4"] Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.699935 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvv6q\" (UniqueName: \"kubernetes.io/projected/33ed389a-fa13-445f-826e-a88c04ef059d-kube-api-access-fvv6q\") pod \"glance-db-create-kplsg\" (UID: \"33ed389a-fa13-445f-826e-a88c04ef059d\") " pod="openstack/glance-db-create-kplsg" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.752660 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmx24\" (UniqueName: \"kubernetes.io/projected/d4632708-aa04-4b74-bdda-99650680eac8-kube-api-access-pmx24\") pod \"glance-c428-account-create-update-x48t4\" (UID: \"d4632708-aa04-4b74-bdda-99650680eac8\") " pod="openstack/glance-c428-account-create-update-x48t4" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.752731 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4632708-aa04-4b74-bdda-99650680eac8-operator-scripts\") pod \"glance-c428-account-create-update-x48t4\" (UID: \"d4632708-aa04-4b74-bdda-99650680eac8\") " pod="openstack/glance-c428-account-create-update-x48t4" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.765947 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-qwb25"] Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.773597 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qwb25" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.804451 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-qwb25"] Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.854210 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4632708-aa04-4b74-bdda-99650680eac8-operator-scripts\") pod \"glance-c428-account-create-update-x48t4\" (UID: \"d4632708-aa04-4b74-bdda-99650680eac8\") " pod="openstack/glance-c428-account-create-update-x48t4" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.854269 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmx24\" (UniqueName: \"kubernetes.io/projected/d4632708-aa04-4b74-bdda-99650680eac8-kube-api-access-pmx24\") pod \"glance-c428-account-create-update-x48t4\" (UID: \"d4632708-aa04-4b74-bdda-99650680eac8\") " pod="openstack/glance-c428-account-create-update-x48t4" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.854497 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-operator-scripts\") pod \"keystone-db-create-qwb25\" (UID: \"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3\") " pod="openstack/keystone-db-create-qwb25" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.854551 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spw9v\" (UniqueName: \"kubernetes.io/projected/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-kube-api-access-spw9v\") pod \"keystone-db-create-qwb25\" (UID: \"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3\") " pod="openstack/keystone-db-create-qwb25" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.855573 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4632708-aa04-4b74-bdda-99650680eac8-operator-scripts\") pod \"glance-c428-account-create-update-x48t4\" (UID: \"d4632708-aa04-4b74-bdda-99650680eac8\") " pod="openstack/glance-c428-account-create-update-x48t4" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.873896 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kplsg" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.953404 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmx24\" (UniqueName: \"kubernetes.io/projected/d4632708-aa04-4b74-bdda-99650680eac8-kube-api-access-pmx24\") pod \"glance-c428-account-create-update-x48t4\" (UID: \"d4632708-aa04-4b74-bdda-99650680eac8\") " pod="openstack/glance-c428-account-create-update-x48t4" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.959912 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-operator-scripts\") pod \"keystone-db-create-qwb25\" (UID: \"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3\") " pod="openstack/keystone-db-create-qwb25" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.959996 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spw9v\" (UniqueName: \"kubernetes.io/projected/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-kube-api-access-spw9v\") pod \"keystone-db-create-qwb25\" (UID: \"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3\") " pod="openstack/keystone-db-create-qwb25" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.960905 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-operator-scripts\") pod \"keystone-db-create-qwb25\" (UID: \"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3\") " pod="openstack/keystone-db-create-qwb25" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.967559 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c428-account-create-update-x48t4" Mar 17 16:01:13 crc kubenswrapper[4767]: I0317 16:01:13.991043 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spw9v\" (UniqueName: \"kubernetes.io/projected/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-kube-api-access-spw9v\") pod \"keystone-db-create-qwb25\" (UID: \"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3\") " pod="openstack/keystone-db-create-qwb25" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.098833 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qwb25" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.136306 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-3aef-account-create-update-x9gj7"] Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.138103 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3aef-account-create-update-x9gj7" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.150794 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.179301 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae4f3fba-7231-451f-a3cd-d560c85212dd-operator-scripts\") pod \"keystone-3aef-account-create-update-x9gj7\" (UID: \"ae4f3fba-7231-451f-a3cd-d560c85212dd\") " pod="openstack/keystone-3aef-account-create-update-x9gj7" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.179692 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz7cx\" (UniqueName: \"kubernetes.io/projected/ae4f3fba-7231-451f-a3cd-d560c85212dd-kube-api-access-gz7cx\") pod \"keystone-3aef-account-create-update-x9gj7\" (UID: \"ae4f3fba-7231-451f-a3cd-d560c85212dd\") " pod="openstack/keystone-3aef-account-create-update-x9gj7" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.182047 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3aef-account-create-update-x9gj7"] Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.211236 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-n9ftr"] Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.214857 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n9ftr" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.282234 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t4zv\" (UniqueName: \"kubernetes.io/projected/f745e6a5-fa5d-4bee-9df5-df119632fbf0-kube-api-access-4t4zv\") pod \"placement-db-create-n9ftr\" (UID: \"f745e6a5-fa5d-4bee-9df5-df119632fbf0\") " pod="openstack/placement-db-create-n9ftr" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.282328 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz7cx\" (UniqueName: \"kubernetes.io/projected/ae4f3fba-7231-451f-a3cd-d560c85212dd-kube-api-access-gz7cx\") pod \"keystone-3aef-account-create-update-x9gj7\" (UID: \"ae4f3fba-7231-451f-a3cd-d560c85212dd\") " pod="openstack/keystone-3aef-account-create-update-x9gj7" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.282357 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f745e6a5-fa5d-4bee-9df5-df119632fbf0-operator-scripts\") pod \"placement-db-create-n9ftr\" (UID: \"f745e6a5-fa5d-4bee-9df5-df119632fbf0\") " pod="openstack/placement-db-create-n9ftr" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.282696 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae4f3fba-7231-451f-a3cd-d560c85212dd-operator-scripts\") pod \"keystone-3aef-account-create-update-x9gj7\" (UID: \"ae4f3fba-7231-451f-a3cd-d560c85212dd\") " pod="openstack/keystone-3aef-account-create-update-x9gj7" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.283692 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae4f3fba-7231-451f-a3cd-d560c85212dd-operator-scripts\") pod \"keystone-3aef-account-create-update-x9gj7\" (UID: \"ae4f3fba-7231-451f-a3cd-d560c85212dd\") " pod="openstack/keystone-3aef-account-create-update-x9gj7" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.299866 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-n9ftr"] Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.315034 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz7cx\" (UniqueName: \"kubernetes.io/projected/ae4f3fba-7231-451f-a3cd-d560c85212dd-kube-api-access-gz7cx\") pod \"keystone-3aef-account-create-update-x9gj7\" (UID: \"ae4f3fba-7231-451f-a3cd-d560c85212dd\") " pod="openstack/keystone-3aef-account-create-update-x9gj7" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.367714 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9758-account-create-update-xjjlg"] Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.369334 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9758-account-create-update-xjjlg" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.372711 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.381858 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9758-account-create-update-xjjlg"] Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.390824 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-operator-scripts\") pod \"placement-9758-account-create-update-xjjlg\" (UID: \"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf\") " pod="openstack/placement-9758-account-create-update-xjjlg" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.390943 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t4zv\" (UniqueName: \"kubernetes.io/projected/f745e6a5-fa5d-4bee-9df5-df119632fbf0-kube-api-access-4t4zv\") pod \"placement-db-create-n9ftr\" (UID: \"f745e6a5-fa5d-4bee-9df5-df119632fbf0\") " pod="openstack/placement-db-create-n9ftr" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.391019 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f745e6a5-fa5d-4bee-9df5-df119632fbf0-operator-scripts\") pod \"placement-db-create-n9ftr\" (UID: \"f745e6a5-fa5d-4bee-9df5-df119632fbf0\") " pod="openstack/placement-db-create-n9ftr" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.391114 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hpmj\" (UniqueName: \"kubernetes.io/projected/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-kube-api-access-9hpmj\") pod \"placement-9758-account-create-update-xjjlg\" (UID: \"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf\") " pod="openstack/placement-9758-account-create-update-xjjlg" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.392424 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f745e6a5-fa5d-4bee-9df5-df119632fbf0-operator-scripts\") pod \"placement-db-create-n9ftr\" (UID: \"f745e6a5-fa5d-4bee-9df5-df119632fbf0\") " pod="openstack/placement-db-create-n9ftr" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.411915 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t4zv\" (UniqueName: \"kubernetes.io/projected/f745e6a5-fa5d-4bee-9df5-df119632fbf0-kube-api-access-4t4zv\") pod \"placement-db-create-n9ftr\" (UID: \"f745e6a5-fa5d-4bee-9df5-df119632fbf0\") " pod="openstack/placement-db-create-n9ftr" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.494480 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-operator-scripts\") pod \"placement-9758-account-create-update-xjjlg\" (UID: \"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf\") " pod="openstack/placement-9758-account-create-update-xjjlg" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.494629 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hpmj\" (UniqueName: \"kubernetes.io/projected/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-kube-api-access-9hpmj\") pod \"placement-9758-account-create-update-xjjlg\" (UID: \"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf\") " pod="openstack/placement-9758-account-create-update-xjjlg" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.496496 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-operator-scripts\") pod \"placement-9758-account-create-update-xjjlg\" (UID: \"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf\") " pod="openstack/placement-9758-account-create-update-xjjlg" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.498985 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3aef-account-create-update-x9gj7" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.513663 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hpmj\" (UniqueName: \"kubernetes.io/projected/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-kube-api-access-9hpmj\") pod \"placement-9758-account-create-update-xjjlg\" (UID: \"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf\") " pod="openstack/placement-9758-account-create-update-xjjlg" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.571144 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n9ftr" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.601537 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.697598 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9758-account-create-update-xjjlg" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.822689 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.823503 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.833462 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 17 16:01:14 crc kubenswrapper[4767]: I0317 16:01:14.833730 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 17 16:01:15 crc kubenswrapper[4767]: I0317 16:01:15.594962 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f","Type":"ContainerStarted","Data":"66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de"} Mar 17 16:01:15 crc kubenswrapper[4767]: I0317 16:01:15.650164 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.459200 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:16 crc kubenswrapper[4767]: E0317 16:01:16.459460 4767 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 16:01:16 crc kubenswrapper[4767]: E0317 16:01:16.459603 4767 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 16:01:16 crc kubenswrapper[4767]: E0317 16:01:16.459691 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift podName:803604b0-bf06-4c1b-bdce-338284abc323 nodeName:}" failed. No retries permitted until 2026-03-17 16:01:24.459667464 +0000 UTC m=+1475.872983511 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift") pod "swift-storage-0" (UID: "803604b0-bf06-4c1b-bdce-338284abc323") : configmap "swift-ring-files" not found Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.570364 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.611201 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.611950 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-vhvs8" event={"ID":"82b61927-c834-4a58-a140-c2c63686f6f1","Type":"ContainerDied","Data":"465f6192f422eac5b7e0151f29891ab6b8ad6b9629d06c1080087fd1eac41f50"} Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.611985 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="465f6192f422eac5b7e0151f29891ab6b8ad6b9629d06c1080087fd1eac41f50" Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.614534 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" event={"ID":"22e12d8e-e55f-4839-afdb-a03c60523aed","Type":"ContainerStarted","Data":"f509fad8be14eb79e9f5479943adce5fcaff6b780313420d51fbe762b4794a7d"} Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.617560 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-z5gph" event={"ID":"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c","Type":"ContainerStarted","Data":"87469601ae350571d0bc7968b18e68646b29ba55c316ac2268f336d71c9e7b2c"} Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.664580 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82b61927-c834-4a58-a140-c2c63686f6f1-operator-scripts\") pod \"82b61927-c834-4a58-a140-c2c63686f6f1\" (UID: \"82b61927-c834-4a58-a140-c2c63686f6f1\") " Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.664957 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cl7c\" (UniqueName: \"kubernetes.io/projected/82b61927-c834-4a58-a140-c2c63686f6f1-kube-api-access-6cl7c\") pod \"82b61927-c834-4a58-a140-c2c63686f6f1\" (UID: \"82b61927-c834-4a58-a140-c2c63686f6f1\") " Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.674183 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82b61927-c834-4a58-a140-c2c63686f6f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "82b61927-c834-4a58-a140-c2c63686f6f1" (UID: "82b61927-c834-4a58-a140-c2c63686f6f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.674598 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82b61927-c834-4a58-a140-c2c63686f6f1-kube-api-access-6cl7c" (OuterVolumeSpecName: "kube-api-access-6cl7c") pod "82b61927-c834-4a58-a140-c2c63686f6f1" (UID: "82b61927-c834-4a58-a140-c2c63686f6f1"). InnerVolumeSpecName "kube-api-access-6cl7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.768690 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cl7c\" (UniqueName: \"kubernetes.io/projected/82b61927-c834-4a58-a140-c2c63686f6f1-kube-api-access-6cl7c\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:16 crc kubenswrapper[4767]: I0317 16:01:16.769041 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82b61927-c834-4a58-a140-c2c63686f6f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:17 crc kubenswrapper[4767]: I0317 16:01:17.636546 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:01:17 crc kubenswrapper[4767]: I0317 16:01:17.640803 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-qwb25"] Mar 17 16:01:17 crc kubenswrapper[4767]: I0317 16:01:17.690097 4767 generic.go:334] "Generic (PLEG): container finished" podID="22e12d8e-e55f-4839-afdb-a03c60523aed" containerID="50a8e92a1f7ae1f117fb00c47296a9dac97c1b41dda1b92092c3d4d4a35f537a" exitCode=0 Mar 17 16:01:17 crc kubenswrapper[4767]: I0317 16:01:17.690244 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" event={"ID":"22e12d8e-e55f-4839-afdb-a03c60523aed","Type":"ContainerDied","Data":"50a8e92a1f7ae1f117fb00c47296a9dac97c1b41dda1b92092c3d4d4a35f537a"} Mar 17 16:01:17 crc kubenswrapper[4767]: W0317 16:01:17.705162 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf70029ee_d0ed_4aaa_8907_f3bfb4b196b3.slice/crio-06cb7258eff0015447af5668f19d739dfef33eb088c3ef76982ca4c82704f93e WatchSource:0}: Error finding container 06cb7258eff0015447af5668f19d739dfef33eb088c3ef76982ca4c82704f93e: Status 404 returned error can't find the container with id 06cb7258eff0015447af5668f19d739dfef33eb088c3ef76982ca4c82704f93e Mar 17 16:01:17 crc kubenswrapper[4767]: I0317 16:01:17.728629 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-cqcdm" event={"ID":"500b62ad-a562-419e-a44a-85030e12bf8b","Type":"ContainerStarted","Data":"c04d1896324154528cb7fccb824fd1e8af7ff5b8cab07299774c184cf2bd0702"} Mar 17 16:01:17 crc kubenswrapper[4767]: I0317 16:01:17.753507 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-z5gph" event={"ID":"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c","Type":"ContainerStarted","Data":"e08a06a020465497aeb9d5f24945601e389e57714d515764ca2d594873a36436"} Mar 17 16:01:17 crc kubenswrapper[4767]: I0317 16:01:17.774921 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-zvjq6"] Mar 17 16:01:17 crc kubenswrapper[4767]: I0317 16:01:17.775246 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" podUID="b2781552-ff04-4f51-8901-7abb87dd1369" containerName="dnsmasq-dns" containerID="cri-o://8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285" gracePeriod=10 Mar 17 16:01:17 crc kubenswrapper[4767]: I0317 16:01:17.775870 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-cqcdm" podStartSLOduration=2.579505971 podStartE2EDuration="8.775842953s" podCreationTimestamp="2026-03-17 16:01:09 +0000 UTC" firstStartedPulling="2026-03-17 16:01:10.404816328 +0000 UTC m=+1461.818132375" lastFinishedPulling="2026-03-17 16:01:16.60115331 +0000 UTC m=+1468.014469357" observedRunningTime="2026-03-17 16:01:17.775763251 +0000 UTC m=+1469.189079298" watchObservedRunningTime="2026-03-17 16:01:17.775842953 +0000 UTC m=+1469.189159000" Mar 17 16:01:17 crc kubenswrapper[4767]: I0317 16:01:17.844774 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9758-account-create-update-xjjlg"] Mar 17 16:01:17 crc kubenswrapper[4767]: W0317 16:01:17.879595 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4632708_aa04_4b74_bdda_99650680eac8.slice/crio-248b48bef1988c8afd94fdfbcfd8165ae92ad620c06d846657df17bd5431e584 WatchSource:0}: Error finding container 248b48bef1988c8afd94fdfbcfd8165ae92ad620c06d846657df17bd5431e584: Status 404 returned error can't find the container with id 248b48bef1988c8afd94fdfbcfd8165ae92ad620c06d846657df17bd5431e584 Mar 17 16:01:17 crc kubenswrapper[4767]: I0317 16:01:17.900855 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c428-account-create-update-x48t4"] Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.060252 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3aef-account-create-update-x9gj7"] Mar 17 16:01:18 crc kubenswrapper[4767]: W0317 16:01:18.065144 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae4f3fba_7231_451f_a3cd_d560c85212dd.slice/crio-3ccb7dd258e252dd1cf5fd4677c25067e0bcfc12010b3d81e1df1513399412c8 WatchSource:0}: Error finding container 3ccb7dd258e252dd1cf5fd4677c25067e0bcfc12010b3d81e1df1513399412c8: Status 404 returned error can't find the container with id 3ccb7dd258e252dd1cf5fd4677c25067e0bcfc12010b3d81e1df1513399412c8 Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.234319 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kplsg"] Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.288236 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-n9ftr"] Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.435596 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.562829 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-config\") pod \"b2781552-ff04-4f51-8901-7abb87dd1369\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.563312 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbxh9\" (UniqueName: \"kubernetes.io/projected/b2781552-ff04-4f51-8901-7abb87dd1369-kube-api-access-wbxh9\") pod \"b2781552-ff04-4f51-8901-7abb87dd1369\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.563375 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-nb\") pod \"b2781552-ff04-4f51-8901-7abb87dd1369\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.563492 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-sb\") pod \"b2781552-ff04-4f51-8901-7abb87dd1369\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.563565 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-dns-svc\") pod \"b2781552-ff04-4f51-8901-7abb87dd1369\" (UID: \"b2781552-ff04-4f51-8901-7abb87dd1369\") " Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.574465 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2781552-ff04-4f51-8901-7abb87dd1369-kube-api-access-wbxh9" (OuterVolumeSpecName: "kube-api-access-wbxh9") pod "b2781552-ff04-4f51-8901-7abb87dd1369" (UID: "b2781552-ff04-4f51-8901-7abb87dd1369"). InnerVolumeSpecName "kube-api-access-wbxh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.656687 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b2781552-ff04-4f51-8901-7abb87dd1369" (UID: "b2781552-ff04-4f51-8901-7abb87dd1369"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.658447 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b2781552-ff04-4f51-8901-7abb87dd1369" (UID: "b2781552-ff04-4f51-8901-7abb87dd1369"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.667394 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.667456 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.667468 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbxh9\" (UniqueName: \"kubernetes.io/projected/b2781552-ff04-4f51-8901-7abb87dd1369-kube-api-access-wbxh9\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.691851 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-config" (OuterVolumeSpecName: "config") pod "b2781552-ff04-4f51-8901-7abb87dd1369" (UID: "b2781552-ff04-4f51-8901-7abb87dd1369"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.706008 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b2781552-ff04-4f51-8901-7abb87dd1369" (UID: "b2781552-ff04-4f51-8901-7abb87dd1369"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.768894 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.768928 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2781552-ff04-4f51-8901-7abb87dd1369-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.772877 4767 generic.go:334] "Generic (PLEG): container finished" podID="85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf" containerID="9b0e569d86178fd324ef0aaf639095da90f4eafa4e2367069824febfa9ecfe9e" exitCode=0 Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.772954 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9758-account-create-update-xjjlg" event={"ID":"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf","Type":"ContainerDied","Data":"9b0e569d86178fd324ef0aaf639095da90f4eafa4e2367069824febfa9ecfe9e"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.772988 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9758-account-create-update-xjjlg" event={"ID":"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf","Type":"ContainerStarted","Data":"4fa795d9bd8859231375d104eef8b5df8a6b70fd6b73e371f223edcac141be7b"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.790423 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-n9ftr" event={"ID":"f745e6a5-fa5d-4bee-9df5-df119632fbf0","Type":"ContainerStarted","Data":"d0d30ee80606f34c1edb9b3aae94618126577c6d6a4159354109d62d1860ac05"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.790478 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-n9ftr" event={"ID":"f745e6a5-fa5d-4bee-9df5-df119632fbf0","Type":"ContainerStarted","Data":"c1c9c6d34ed378fc914fc21f9f4102e80b616c81ae16747f160227d60e624a99"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.794754 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3aef-account-create-update-x9gj7" event={"ID":"ae4f3fba-7231-451f-a3cd-d560c85212dd","Type":"ContainerStarted","Data":"25a5ddb7324c160d2823c76c7b6b889aded1ecf0cb07c04712f1d7c75412d7f4"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.794809 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3aef-account-create-update-x9gj7" event={"ID":"ae4f3fba-7231-451f-a3cd-d560c85212dd","Type":"ContainerStarted","Data":"3ccb7dd258e252dd1cf5fd4677c25067e0bcfc12010b3d81e1df1513399412c8"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.797576 4767 generic.go:334] "Generic (PLEG): container finished" podID="7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c" containerID="e08a06a020465497aeb9d5f24945601e389e57714d515764ca2d594873a36436" exitCode=0 Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.799813 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-z5gph" event={"ID":"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c","Type":"ContainerDied","Data":"e08a06a020465497aeb9d5f24945601e389e57714d515764ca2d594873a36436"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.805601 4767 generic.go:334] "Generic (PLEG): container finished" podID="f70029ee-d0ed-4aaa-8907-f3bfb4b196b3" containerID="5c726a1f548330a53788b2877ffd12622841da12ca66e91210410e7b3d99fe87" exitCode=0 Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.805722 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qwb25" event={"ID":"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3","Type":"ContainerDied","Data":"5c726a1f548330a53788b2877ffd12622841da12ca66e91210410e7b3d99fe87"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.805816 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qwb25" event={"ID":"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3","Type":"ContainerStarted","Data":"06cb7258eff0015447af5668f19d739dfef33eb088c3ef76982ca4c82704f93e"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.809114 4767 generic.go:334] "Generic (PLEG): container finished" podID="b2781552-ff04-4f51-8901-7abb87dd1369" containerID="8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285" exitCode=0 Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.809273 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.809298 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" event={"ID":"b2781552-ff04-4f51-8901-7abb87dd1369","Type":"ContainerDied","Data":"8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.809361 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-zvjq6" event={"ID":"b2781552-ff04-4f51-8901-7abb87dd1369","Type":"ContainerDied","Data":"6d2594ab02f6bf4e993b5fb971d163b9bcb627d76ae0805eee84fd9b153e650f"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.809388 4767 scope.go:117] "RemoveContainer" containerID="8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.816215 4767 generic.go:334] "Generic (PLEG): container finished" podID="d4632708-aa04-4b74-bdda-99650680eac8" containerID="5a541f5e86b1f3f83a9bd446147a081f76ea3343987be804508edd1004ad4db8" exitCode=0 Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.816451 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c428-account-create-update-x48t4" event={"ID":"d4632708-aa04-4b74-bdda-99650680eac8","Type":"ContainerDied","Data":"5a541f5e86b1f3f83a9bd446147a081f76ea3343987be804508edd1004ad4db8"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.816488 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c428-account-create-update-x48t4" event={"ID":"d4632708-aa04-4b74-bdda-99650680eac8","Type":"ContainerStarted","Data":"248b48bef1988c8afd94fdfbcfd8165ae92ad620c06d846657df17bd5431e584"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.828145 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-3aef-account-create-update-x9gj7" podStartSLOduration=4.828123577 podStartE2EDuration="4.828123577s" podCreationTimestamp="2026-03-17 16:01:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:01:18.822360374 +0000 UTC m=+1470.235676441" watchObservedRunningTime="2026-03-17 16:01:18.828123577 +0000 UTC m=+1470.241439624" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.832257 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kplsg" event={"ID":"33ed389a-fa13-445f-826e-a88c04ef059d","Type":"ContainerStarted","Data":"656cb70f5661b543ee3d2038ffd65436446a83cee9f1d0a01b17533d30f9b31e"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.832343 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kplsg" event={"ID":"33ed389a-fa13-445f-826e-a88c04ef059d","Type":"ContainerStarted","Data":"4aef7ee855d7a13bfddbc3f03819a14698136ef0f799af80a0330c74cade7f75"} Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.856162 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-n9ftr" podStartSLOduration=4.856131045 podStartE2EDuration="4.856131045s" podCreationTimestamp="2026-03-17 16:01:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:01:18.838898415 +0000 UTC m=+1470.252214472" watchObservedRunningTime="2026-03-17 16:01:18.856131045 +0000 UTC m=+1470.269447092" Mar 17 16:01:18 crc kubenswrapper[4767]: I0317 16:01:18.900054 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-kplsg" podStartSLOduration=5.9000008059999995 podStartE2EDuration="5.900000806s" podCreationTimestamp="2026-03-17 16:01:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:01:18.891966092 +0000 UTC m=+1470.305282159" watchObservedRunningTime="2026-03-17 16:01:18.900000806 +0000 UTC m=+1470.313316853" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.195087 4767 scope.go:117] "RemoveContainer" containerID="0bd89e5257dfbc11b4b35aafb16fd4e99dcbf69ba863634b70c351baf8ab7303" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.195523 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-zvjq6"] Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.202287 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-zvjq6"] Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.233910 4767 scope.go:117] "RemoveContainer" containerID="8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285" Mar 17 16:01:19 crc kubenswrapper[4767]: E0317 16:01:19.234647 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285\": container with ID starting with 8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285 not found: ID does not exist" containerID="8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.234720 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285"} err="failed to get container status \"8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285\": rpc error: code = NotFound desc = could not find container \"8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285\": container with ID starting with 8b19adfb7c1af65c684b7dcdedba5cf7b90ef1780c65ee88d0bdf555939a9285 not found: ID does not exist" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.234756 4767 scope.go:117] "RemoveContainer" containerID="0bd89e5257dfbc11b4b35aafb16fd4e99dcbf69ba863634b70c351baf8ab7303" Mar 17 16:01:19 crc kubenswrapper[4767]: E0317 16:01:19.235321 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bd89e5257dfbc11b4b35aafb16fd4e99dcbf69ba863634b70c351baf8ab7303\": container with ID starting with 0bd89e5257dfbc11b4b35aafb16fd4e99dcbf69ba863634b70c351baf8ab7303 not found: ID does not exist" containerID="0bd89e5257dfbc11b4b35aafb16fd4e99dcbf69ba863634b70c351baf8ab7303" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.235410 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bd89e5257dfbc11b4b35aafb16fd4e99dcbf69ba863634b70c351baf8ab7303"} err="failed to get container status \"0bd89e5257dfbc11b4b35aafb16fd4e99dcbf69ba863634b70c351baf8ab7303\": rpc error: code = NotFound desc = could not find container \"0bd89e5257dfbc11b4b35aafb16fd4e99dcbf69ba863634b70c351baf8ab7303\": container with ID starting with 0bd89e5257dfbc11b4b35aafb16fd4e99dcbf69ba863634b70c351baf8ab7303 not found: ID does not exist" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.348403 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-z5gph" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.397535 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-operator-scripts\") pod \"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c\" (UID: \"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c\") " Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.397691 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2s6g\" (UniqueName: \"kubernetes.io/projected/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-kube-api-access-l2s6g\") pod \"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c\" (UID: \"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c\") " Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.398617 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c" (UID: "7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.408524 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-kube-api-access-l2s6g" (OuterVolumeSpecName: "kube-api-access-l2s6g") pod "7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c" (UID: "7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c"). InnerVolumeSpecName "kube-api-access-l2s6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.424746 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2781552-ff04-4f51-8901-7abb87dd1369" path="/var/lib/kubelet/pods/b2781552-ff04-4f51-8901-7abb87dd1369/volumes" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.501006 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.501055 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2s6g\" (UniqueName: \"kubernetes.io/projected/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c-kube-api-access-l2s6g\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.571493 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.705512 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22e12d8e-e55f-4839-afdb-a03c60523aed-operator-scripts\") pod \"22e12d8e-e55f-4839-afdb-a03c60523aed\" (UID: \"22e12d8e-e55f-4839-afdb-a03c60523aed\") " Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.705638 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6gm5\" (UniqueName: \"kubernetes.io/projected/22e12d8e-e55f-4839-afdb-a03c60523aed-kube-api-access-l6gm5\") pod \"22e12d8e-e55f-4839-afdb-a03c60523aed\" (UID: \"22e12d8e-e55f-4839-afdb-a03c60523aed\") " Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.706024 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22e12d8e-e55f-4839-afdb-a03c60523aed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "22e12d8e-e55f-4839-afdb-a03c60523aed" (UID: "22e12d8e-e55f-4839-afdb-a03c60523aed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.706299 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22e12d8e-e55f-4839-afdb-a03c60523aed-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.726665 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22e12d8e-e55f-4839-afdb-a03c60523aed-kube-api-access-l6gm5" (OuterVolumeSpecName: "kube-api-access-l6gm5") pod "22e12d8e-e55f-4839-afdb-a03c60523aed" (UID: "22e12d8e-e55f-4839-afdb-a03c60523aed"). InnerVolumeSpecName "kube-api-access-l6gm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.808408 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6gm5\" (UniqueName: \"kubernetes.io/projected/22e12d8e-e55f-4839-afdb-a03c60523aed-kube-api-access-l6gm5\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.876644 4767 generic.go:334] "Generic (PLEG): container finished" podID="33ed389a-fa13-445f-826e-a88c04ef059d" containerID="656cb70f5661b543ee3d2038ffd65436446a83cee9f1d0a01b17533d30f9b31e" exitCode=0 Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.877057 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kplsg" event={"ID":"33ed389a-fa13-445f-826e-a88c04ef059d","Type":"ContainerDied","Data":"656cb70f5661b543ee3d2038ffd65436446a83cee9f1d0a01b17533d30f9b31e"} Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.884882 4767 generic.go:334] "Generic (PLEG): container finished" podID="f745e6a5-fa5d-4bee-9df5-df119632fbf0" containerID="d0d30ee80606f34c1edb9b3aae94618126577c6d6a4159354109d62d1860ac05" exitCode=0 Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.884977 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-n9ftr" event={"ID":"f745e6a5-fa5d-4bee-9df5-df119632fbf0","Type":"ContainerDied","Data":"d0d30ee80606f34c1edb9b3aae94618126577c6d6a4159354109d62d1860ac05"} Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.890617 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" event={"ID":"22e12d8e-e55f-4839-afdb-a03c60523aed","Type":"ContainerDied","Data":"f509fad8be14eb79e9f5479943adce5fcaff6b780313420d51fbe762b4794a7d"} Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.890672 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-ccf0-account-create-update-vzhjl" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.890696 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f509fad8be14eb79e9f5479943adce5fcaff6b780313420d51fbe762b4794a7d" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.895236 4767 generic.go:334] "Generic (PLEG): container finished" podID="ae4f3fba-7231-451f-a3cd-d560c85212dd" containerID="25a5ddb7324c160d2823c76c7b6b889aded1ecf0cb07c04712f1d7c75412d7f4" exitCode=0 Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.895315 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3aef-account-create-update-x9gj7" event={"ID":"ae4f3fba-7231-451f-a3cd-d560c85212dd","Type":"ContainerDied","Data":"25a5ddb7324c160d2823c76c7b6b889aded1ecf0cb07c04712f1d7c75412d7f4"} Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.904096 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-z5gph" Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.904741 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-z5gph" event={"ID":"7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c","Type":"ContainerDied","Data":"87469601ae350571d0bc7968b18e68646b29ba55c316ac2268f336d71c9e7b2c"} Mar 17 16:01:19 crc kubenswrapper[4767]: I0317 16:01:19.904920 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87469601ae350571d0bc7968b18e68646b29ba55c316ac2268f336d71c9e7b2c" Mar 17 16:01:20 crc kubenswrapper[4767]: E0317 16:01:20.141573 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22e12d8e_e55f_4839_afdb_a03c60523aed.slice/crio-f509fad8be14eb79e9f5479943adce5fcaff6b780313420d51fbe762b4794a7d\": RecentStats: unable to find data in memory cache]" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.442524 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c428-account-create-update-x48t4" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.546789 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmx24\" (UniqueName: \"kubernetes.io/projected/d4632708-aa04-4b74-bdda-99650680eac8-kube-api-access-pmx24\") pod \"d4632708-aa04-4b74-bdda-99650680eac8\" (UID: \"d4632708-aa04-4b74-bdda-99650680eac8\") " Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.547078 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4632708-aa04-4b74-bdda-99650680eac8-operator-scripts\") pod \"d4632708-aa04-4b74-bdda-99650680eac8\" (UID: \"d4632708-aa04-4b74-bdda-99650680eac8\") " Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.548973 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4632708-aa04-4b74-bdda-99650680eac8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d4632708-aa04-4b74-bdda-99650680eac8" (UID: "d4632708-aa04-4b74-bdda-99650680eac8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.553372 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4632708-aa04-4b74-bdda-99650680eac8-kube-api-access-pmx24" (OuterVolumeSpecName: "kube-api-access-pmx24") pod "d4632708-aa04-4b74-bdda-99650680eac8" (UID: "d4632708-aa04-4b74-bdda-99650680eac8"). InnerVolumeSpecName "kube-api-access-pmx24". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.650907 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmx24\" (UniqueName: \"kubernetes.io/projected/d4632708-aa04-4b74-bdda-99650680eac8-kube-api-access-pmx24\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.651250 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4632708-aa04-4b74-bdda-99650680eac8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.655691 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qwb25" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.671425 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9758-account-create-update-xjjlg" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.752838 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-operator-scripts\") pod \"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3\" (UID: \"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3\") " Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.752931 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hpmj\" (UniqueName: \"kubernetes.io/projected/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-kube-api-access-9hpmj\") pod \"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf\" (UID: \"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf\") " Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.752974 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-operator-scripts\") pod \"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf\" (UID: \"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf\") " Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.753640 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf" (UID: "85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.753645 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f70029ee-d0ed-4aaa-8907-f3bfb4b196b3" (UID: "f70029ee-d0ed-4aaa-8907-f3bfb4b196b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.753841 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spw9v\" (UniqueName: \"kubernetes.io/projected/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-kube-api-access-spw9v\") pod \"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3\" (UID: \"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3\") " Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.755196 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.755233 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.759517 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-kube-api-access-9hpmj" (OuterVolumeSpecName: "kube-api-access-9hpmj") pod "85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf" (UID: "85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf"). InnerVolumeSpecName "kube-api-access-9hpmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.765581 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-kube-api-access-spw9v" (OuterVolumeSpecName: "kube-api-access-spw9v") pod "f70029ee-d0ed-4aaa-8907-f3bfb4b196b3" (UID: "f70029ee-d0ed-4aaa-8907-f3bfb4b196b3"). InnerVolumeSpecName "kube-api-access-spw9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.834419 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.857401 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spw9v\" (UniqueName: \"kubernetes.io/projected/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3-kube-api-access-spw9v\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.857452 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hpmj\" (UniqueName: \"kubernetes.io/projected/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf-kube-api-access-9hpmj\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.965264 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c428-account-create-update-x48t4" event={"ID":"d4632708-aa04-4b74-bdda-99650680eac8","Type":"ContainerDied","Data":"248b48bef1988c8afd94fdfbcfd8165ae92ad620c06d846657df17bd5431e584"} Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.965646 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="248b48bef1988c8afd94fdfbcfd8165ae92ad620c06d846657df17bd5431e584" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.965793 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c428-account-create-update-x48t4" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.998474 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9758-account-create-update-xjjlg" event={"ID":"85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf","Type":"ContainerDied","Data":"4fa795d9bd8859231375d104eef8b5df8a6b70fd6b73e371f223edcac141be7b"} Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.998519 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fa795d9bd8859231375d104eef8b5df8a6b70fd6b73e371f223edcac141be7b" Mar 17 16:01:20 crc kubenswrapper[4767]: I0317 16:01:20.998588 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9758-account-create-update-xjjlg" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.002239 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qwb25" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.002925 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qwb25" event={"ID":"f70029ee-d0ed-4aaa-8907-f3bfb4b196b3","Type":"ContainerDied","Data":"06cb7258eff0015447af5668f19d739dfef33eb088c3ef76982ca4c82704f93e"} Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.002976 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06cb7258eff0015447af5668f19d739dfef33eb088c3ef76982ca4c82704f93e" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.035230 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 17 16:01:21 crc kubenswrapper[4767]: E0317 16:01:21.035773 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.035787 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: E0317 16:01:21.035801 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4632708-aa04-4b74-bdda-99650680eac8" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.035806 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4632708-aa04-4b74-bdda-99650680eac8" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: E0317 16:01:21.035824 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2781552-ff04-4f51-8901-7abb87dd1369" containerName="dnsmasq-dns" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.035831 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2781552-ff04-4f51-8901-7abb87dd1369" containerName="dnsmasq-dns" Mar 17 16:01:21 crc kubenswrapper[4767]: E0317 16:01:21.035845 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2781552-ff04-4f51-8901-7abb87dd1369" containerName="init" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.035851 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2781552-ff04-4f51-8901-7abb87dd1369" containerName="init" Mar 17 16:01:21 crc kubenswrapper[4767]: E0317 16:01:21.035879 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22e12d8e-e55f-4839-afdb-a03c60523aed" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.035885 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="22e12d8e-e55f-4839-afdb-a03c60523aed" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: E0317 16:01:21.035895 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b61927-c834-4a58-a140-c2c63686f6f1" containerName="mariadb-database-create" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.035901 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b61927-c834-4a58-a140-c2c63686f6f1" containerName="mariadb-database-create" Mar 17 16:01:21 crc kubenswrapper[4767]: E0317 16:01:21.035913 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.035919 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: E0317 16:01:21.035930 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f70029ee-d0ed-4aaa-8907-f3bfb4b196b3" containerName="mariadb-database-create" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.035936 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f70029ee-d0ed-4aaa-8907-f3bfb4b196b3" containerName="mariadb-database-create" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.036161 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.036274 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="22e12d8e-e55f-4839-afdb-a03c60523aed" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.036307 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4632708-aa04-4b74-bdda-99650680eac8" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.036316 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="82b61927-c834-4a58-a140-c2c63686f6f1" containerName="mariadb-database-create" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.036327 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f70029ee-d0ed-4aaa-8907-f3bfb4b196b3" containerName="mariadb-database-create" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.036336 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c" containerName="mariadb-account-create-update" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.036345 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2781552-ff04-4f51-8901-7abb87dd1369" containerName="dnsmasq-dns" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.038274 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.045832 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-spdxm" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.046029 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.046146 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.046446 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.080945 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2d93f34-029a-4de9-941f-e04a7c957fc3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.081423 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2d93f34-029a-4de9-941f-e04a7c957fc3-config\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.081529 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2d93f34-029a-4de9-941f-e04a7c957fc3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.081552 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2xm7\" (UniqueName: \"kubernetes.io/projected/e2d93f34-029a-4de9-941f-e04a7c957fc3-kube-api-access-s2xm7\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.081956 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d93f34-029a-4de9-941f-e04a7c957fc3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.082051 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2d93f34-029a-4de9-941f-e04a7c957fc3-scripts\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.082088 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e2d93f34-029a-4de9-941f-e04a7c957fc3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.090354 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.186514 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2d93f34-029a-4de9-941f-e04a7c957fc3-config\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.186613 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2d93f34-029a-4de9-941f-e04a7c957fc3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.186647 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2xm7\" (UniqueName: \"kubernetes.io/projected/e2d93f34-029a-4de9-941f-e04a7c957fc3-kube-api-access-s2xm7\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.186818 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d93f34-029a-4de9-941f-e04a7c957fc3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.186882 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2d93f34-029a-4de9-941f-e04a7c957fc3-scripts\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.186921 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e2d93f34-029a-4de9-941f-e04a7c957fc3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.186982 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2d93f34-029a-4de9-941f-e04a7c957fc3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.187606 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2d93f34-029a-4de9-941f-e04a7c957fc3-config\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.190494 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e2d93f34-029a-4de9-941f-e04a7c957fc3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.190677 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2d93f34-029a-4de9-941f-e04a7c957fc3-scripts\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.193839 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2d93f34-029a-4de9-941f-e04a7c957fc3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.193969 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d93f34-029a-4de9-941f-e04a7c957fc3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.199993 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2d93f34-029a-4de9-941f-e04a7c957fc3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.210812 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2xm7\" (UniqueName: \"kubernetes.io/projected/e2d93f34-029a-4de9-941f-e04a7c957fc3-kube-api-access-s2xm7\") pod \"ovn-northd-0\" (UID: \"e2d93f34-029a-4de9-941f-e04a7c957fc3\") " pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.388073 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.952935 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f"] Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.954822 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" Mar 17 16:01:21 crc kubenswrapper[4767]: I0317 16:01:21.965280 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f"] Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.020803 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghdbc\" (UniqueName: \"kubernetes.io/projected/e718a8e4-5847-4051-a7d3-d01f0c22d801-kube-api-access-ghdbc\") pod \"mysqld-exporter-openstack-cell1-db-create-w9n2f\" (UID: \"e718a8e4-5847-4051-a7d3-d01f0c22d801\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.020855 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e718a8e4-5847-4051-a7d3-d01f0c22d801-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-w9n2f\" (UID: \"e718a8e4-5847-4051-a7d3-d01f0c22d801\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.123553 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghdbc\" (UniqueName: \"kubernetes.io/projected/e718a8e4-5847-4051-a7d3-d01f0c22d801-kube-api-access-ghdbc\") pod \"mysqld-exporter-openstack-cell1-db-create-w9n2f\" (UID: \"e718a8e4-5847-4051-a7d3-d01f0c22d801\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.123626 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e718a8e4-5847-4051-a7d3-d01f0c22d801-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-w9n2f\" (UID: \"e718a8e4-5847-4051-a7d3-d01f0c22d801\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.124563 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e718a8e4-5847-4051-a7d3-d01f0c22d801-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-w9n2f\" (UID: \"e718a8e4-5847-4051-a7d3-d01f0c22d801\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.157362 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghdbc\" (UniqueName: \"kubernetes.io/projected/e718a8e4-5847-4051-a7d3-d01f0c22d801-kube-api-access-ghdbc\") pod \"mysqld-exporter-openstack-cell1-db-create-w9n2f\" (UID: \"e718a8e4-5847-4051-a7d3-d01f0c22d801\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.299830 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.336090 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-fa1b-account-create-update-nkxnj"] Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.374681 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.381352 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.454819 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-fa1b-account-create-update-nkxnj"] Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.466132 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ctns\" (UniqueName: \"kubernetes.io/projected/89f09b66-7eec-4ad8-9116-83fa9d80fead-kube-api-access-8ctns\") pod \"mysqld-exporter-fa1b-account-create-update-nkxnj\" (UID: \"89f09b66-7eec-4ad8-9116-83fa9d80fead\") " pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.466344 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89f09b66-7eec-4ad8-9116-83fa9d80fead-operator-scripts\") pod \"mysqld-exporter-fa1b-account-create-update-nkxnj\" (UID: \"89f09b66-7eec-4ad8-9116-83fa9d80fead\") " pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.492355 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-z5gph"] Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.506486 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-z5gph"] Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.569513 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89f09b66-7eec-4ad8-9116-83fa9d80fead-operator-scripts\") pod \"mysqld-exporter-fa1b-account-create-update-nkxnj\" (UID: \"89f09b66-7eec-4ad8-9116-83fa9d80fead\") " pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.569745 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ctns\" (UniqueName: \"kubernetes.io/projected/89f09b66-7eec-4ad8-9116-83fa9d80fead-kube-api-access-8ctns\") pod \"mysqld-exporter-fa1b-account-create-update-nkxnj\" (UID: \"89f09b66-7eec-4ad8-9116-83fa9d80fead\") " pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.570890 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89f09b66-7eec-4ad8-9116-83fa9d80fead-operator-scripts\") pod \"mysqld-exporter-fa1b-account-create-update-nkxnj\" (UID: \"89f09b66-7eec-4ad8-9116-83fa9d80fead\") " pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.608431 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ctns\" (UniqueName: \"kubernetes.io/projected/89f09b66-7eec-4ad8-9116-83fa9d80fead-kube-api-access-8ctns\") pod \"mysqld-exporter-fa1b-account-create-update-nkxnj\" (UID: \"89f09b66-7eec-4ad8-9116-83fa9d80fead\") " pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" Mar 17 16:01:22 crc kubenswrapper[4767]: I0317 16:01:22.785317 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" Mar 17 16:01:23 crc kubenswrapper[4767]: I0317 16:01:23.367502 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c" path="/var/lib/kubelet/pods/7ead8a5c-fbc0-41a0-adef-0e2f9aa20d9c/volumes" Mar 17 16:01:24 crc kubenswrapper[4767]: I0317 16:01:24.514603 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:24 crc kubenswrapper[4767]: E0317 16:01:24.514877 4767 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 16:01:24 crc kubenswrapper[4767]: E0317 16:01:24.515125 4767 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 16:01:24 crc kubenswrapper[4767]: E0317 16:01:24.515196 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift podName:803604b0-bf06-4c1b-bdce-338284abc323 nodeName:}" failed. No retries permitted until 2026-03-17 16:01:40.515165972 +0000 UTC m=+1491.928482019 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift") pod "swift-storage-0" (UID: "803604b0-bf06-4c1b-bdce-338284abc323") : configmap "swift-ring-files" not found Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.061202 4767 generic.go:334] "Generic (PLEG): container finished" podID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" containerID="db91577d9d962897e9869c843454246c300eb4ac87d6b82c10120049bf6d2916" exitCode=0 Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.061307 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e","Type":"ContainerDied","Data":"db91577d9d962897e9869c843454246c300eb4ac87d6b82c10120049bf6d2916"} Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.065798 4767 generic.go:334] "Generic (PLEG): container finished" podID="048fa918-95b2-404e-81f5-8989ed47ec0a" containerID="90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9" exitCode=0 Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.065867 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"048fa918-95b2-404e-81f5-8989ed47ec0a","Type":"ContainerDied","Data":"90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9"} Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.732774 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n9ftr" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.781973 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3aef-account-create-update-x9gj7" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.803727 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kplsg" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.863212 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz7cx\" (UniqueName: \"kubernetes.io/projected/ae4f3fba-7231-451f-a3cd-d560c85212dd-kube-api-access-gz7cx\") pod \"ae4f3fba-7231-451f-a3cd-d560c85212dd\" (UID: \"ae4f3fba-7231-451f-a3cd-d560c85212dd\") " Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.863292 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t4zv\" (UniqueName: \"kubernetes.io/projected/f745e6a5-fa5d-4bee-9df5-df119632fbf0-kube-api-access-4t4zv\") pod \"f745e6a5-fa5d-4bee-9df5-df119632fbf0\" (UID: \"f745e6a5-fa5d-4bee-9df5-df119632fbf0\") " Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.863327 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f745e6a5-fa5d-4bee-9df5-df119632fbf0-operator-scripts\") pod \"f745e6a5-fa5d-4bee-9df5-df119632fbf0\" (UID: \"f745e6a5-fa5d-4bee-9df5-df119632fbf0\") " Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.863610 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae4f3fba-7231-451f-a3cd-d560c85212dd-operator-scripts\") pod \"ae4f3fba-7231-451f-a3cd-d560c85212dd\" (UID: \"ae4f3fba-7231-451f-a3cd-d560c85212dd\") " Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.873379 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f745e6a5-fa5d-4bee-9df5-df119632fbf0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f745e6a5-fa5d-4bee-9df5-df119632fbf0" (UID: "f745e6a5-fa5d-4bee-9df5-df119632fbf0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.877022 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae4f3fba-7231-451f-a3cd-d560c85212dd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae4f3fba-7231-451f-a3cd-d560c85212dd" (UID: "ae4f3fba-7231-451f-a3cd-d560c85212dd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.914650 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f745e6a5-fa5d-4bee-9df5-df119632fbf0-kube-api-access-4t4zv" (OuterVolumeSpecName: "kube-api-access-4t4zv") pod "f745e6a5-fa5d-4bee-9df5-df119632fbf0" (UID: "f745e6a5-fa5d-4bee-9df5-df119632fbf0"). InnerVolumeSpecName "kube-api-access-4t4zv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.921499 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae4f3fba-7231-451f-a3cd-d560c85212dd-kube-api-access-gz7cx" (OuterVolumeSpecName: "kube-api-access-gz7cx") pod "ae4f3fba-7231-451f-a3cd-d560c85212dd" (UID: "ae4f3fba-7231-451f-a3cd-d560c85212dd"). InnerVolumeSpecName "kube-api-access-gz7cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.969801 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvv6q\" (UniqueName: \"kubernetes.io/projected/33ed389a-fa13-445f-826e-a88c04ef059d-kube-api-access-fvv6q\") pod \"33ed389a-fa13-445f-826e-a88c04ef059d\" (UID: \"33ed389a-fa13-445f-826e-a88c04ef059d\") " Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.970032 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33ed389a-fa13-445f-826e-a88c04ef059d-operator-scripts\") pod \"33ed389a-fa13-445f-826e-a88c04ef059d\" (UID: \"33ed389a-fa13-445f-826e-a88c04ef059d\") " Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.970649 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz7cx\" (UniqueName: \"kubernetes.io/projected/ae4f3fba-7231-451f-a3cd-d560c85212dd-kube-api-access-gz7cx\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.970668 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f745e6a5-fa5d-4bee-9df5-df119632fbf0-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.970677 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t4zv\" (UniqueName: \"kubernetes.io/projected/f745e6a5-fa5d-4bee-9df5-df119632fbf0-kube-api-access-4t4zv\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.970687 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae4f3fba-7231-451f-a3cd-d560c85212dd-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.971118 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33ed389a-fa13-445f-826e-a88c04ef059d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "33ed389a-fa13-445f-826e-a88c04ef059d" (UID: "33ed389a-fa13-445f-826e-a88c04ef059d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:25 crc kubenswrapper[4767]: I0317 16:01:25.998929 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33ed389a-fa13-445f-826e-a88c04ef059d-kube-api-access-fvv6q" (OuterVolumeSpecName: "kube-api-access-fvv6q") pod "33ed389a-fa13-445f-826e-a88c04ef059d" (UID: "33ed389a-fa13-445f-826e-a88c04ef059d"). InnerVolumeSpecName "kube-api-access-fvv6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.072923 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvv6q\" (UniqueName: \"kubernetes.io/projected/33ed389a-fa13-445f-826e-a88c04ef059d-kube-api-access-fvv6q\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.072972 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33ed389a-fa13-445f-826e-a88c04ef059d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.080534 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f","Type":"ContainerStarted","Data":"f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05"} Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.090044 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kplsg" event={"ID":"33ed389a-fa13-445f-826e-a88c04ef059d","Type":"ContainerDied","Data":"4aef7ee855d7a13bfddbc3f03819a14698136ef0f799af80a0330c74cade7f75"} Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.090081 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kplsg" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.090086 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4aef7ee855d7a13bfddbc3f03819a14698136ef0f799af80a0330c74cade7f75" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.094013 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-n9ftr" event={"ID":"f745e6a5-fa5d-4bee-9df5-df119632fbf0","Type":"ContainerDied","Data":"c1c9c6d34ed378fc914fc21f9f4102e80b616c81ae16747f160227d60e624a99"} Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.094041 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1c9c6d34ed378fc914fc21f9f4102e80b616c81ae16747f160227d60e624a99" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.094151 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n9ftr" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.098049 4767 generic.go:334] "Generic (PLEG): container finished" podID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" containerID="bfb2e2fa71eea606d992e9bc0235bcb0151201ab641f392b4e91ac2749bdc185" exitCode=0 Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.098140 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"3ad3d357-572b-4b20-bfe6-4aa0b3513d41","Type":"ContainerDied","Data":"bfb2e2fa71eea606d992e9bc0235bcb0151201ab641f392b4e91ac2749bdc185"} Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.109440 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e","Type":"ContainerStarted","Data":"143c311e23cdc0a92bff7b528ae3f0e08fa41bb7556bfc2c67cccbe9c57c1306"} Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.110042 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.119456 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3aef-account-create-update-x9gj7" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.119992 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3aef-account-create-update-x9gj7" event={"ID":"ae4f3fba-7231-451f-a3cd-d560c85212dd","Type":"ContainerDied","Data":"3ccb7dd258e252dd1cf5fd4677c25067e0bcfc12010b3d81e1df1513399412c8"} Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.120054 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ccb7dd258e252dd1cf5fd4677c25067e0bcfc12010b3d81e1df1513399412c8" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.135980 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"048fa918-95b2-404e-81f5-8989ed47ec0a","Type":"ContainerStarted","Data":"b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74"} Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.136936 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.141784 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=7.430531363 podStartE2EDuration="1m27.14174619s" podCreationTimestamp="2026-03-17 15:59:59 +0000 UTC" firstStartedPulling="2026-03-17 16:00:05.953419715 +0000 UTC m=+1397.366735762" lastFinishedPulling="2026-03-17 16:01:25.664634542 +0000 UTC m=+1477.077950589" observedRunningTime="2026-03-17 16:01:26.129483452 +0000 UTC m=+1477.542799509" watchObservedRunningTime="2026-03-17 16:01:26.14174619 +0000 UTC m=+1477.555062237" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.268714 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=44.594471626 podStartE2EDuration="1m39.268680179s" podCreationTimestamp="2026-03-17 15:59:47 +0000 UTC" firstStartedPulling="2026-03-17 15:59:53.468785688 +0000 UTC m=+1384.882101735" lastFinishedPulling="2026-03-17 16:00:48.142994241 +0000 UTC m=+1439.556310288" observedRunningTime="2026-03-17 16:01:26.234672541 +0000 UTC m=+1477.647988598" watchObservedRunningTime="2026-03-17 16:01:26.268680179 +0000 UTC m=+1477.681996236" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.286941 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=42.564440074 podStartE2EDuration="1m40.286911226s" podCreationTimestamp="2026-03-17 15:59:46 +0000 UTC" firstStartedPulling="2026-03-17 15:59:51.919102029 +0000 UTC m=+1383.332418076" lastFinishedPulling="2026-03-17 16:00:49.641573191 +0000 UTC m=+1441.054889228" observedRunningTime="2026-03-17 16:01:26.274055812 +0000 UTC m=+1477.687371869" watchObservedRunningTime="2026-03-17 16:01:26.286911226 +0000 UTC m=+1477.700227293" Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.298260 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-fa1b-account-create-update-nkxnj"] Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.333401 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f"] Mar 17 16:01:26 crc kubenswrapper[4767]: I0317 16:01:26.476485 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.153945 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"3ad3d357-572b-4b20-bfe6-4aa0b3513d41","Type":"ContainerStarted","Data":"2f570a2d5d87e91310496d39d3f5743a020b78ef2da12089bcec9dc84e6c63c1"} Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.154558 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-1" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.158058 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e2d93f34-029a-4de9-941f-e04a7c957fc3","Type":"ContainerStarted","Data":"9566114cb987e3c140767e74fb8990522cd36b6eb5ae3d48b137593668b39752"} Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.161100 4767 generic.go:334] "Generic (PLEG): container finished" podID="e718a8e4-5847-4051-a7d3-d01f0c22d801" containerID="97d3145338f2803a261d84bfbb65c532f444c029b91eb8768f958fe6ee461023" exitCode=0 Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.161190 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" event={"ID":"e718a8e4-5847-4051-a7d3-d01f0c22d801","Type":"ContainerDied","Data":"97d3145338f2803a261d84bfbb65c532f444c029b91eb8768f958fe6ee461023"} Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.161216 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" event={"ID":"e718a8e4-5847-4051-a7d3-d01f0c22d801","Type":"ContainerStarted","Data":"de89fcdeac5695b7e0d912533d08bed9f10405695969442d878ce1406ed3427a"} Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.163086 4767 generic.go:334] "Generic (PLEG): container finished" podID="89f09b66-7eec-4ad8-9116-83fa9d80fead" containerID="4da7550950abbb01ac3501a84e75b92ab14da305bd43c38473c9eedeed31c059" exitCode=0 Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.164244 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" event={"ID":"89f09b66-7eec-4ad8-9116-83fa9d80fead","Type":"ContainerDied","Data":"4da7550950abbb01ac3501a84e75b92ab14da305bd43c38473c9eedeed31c059"} Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.164268 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" event={"ID":"89f09b66-7eec-4ad8-9116-83fa9d80fead","Type":"ContainerStarted","Data":"a73da47cfb9e20a900b769ea12ce1982e22a749525498dac7263ef08043273b6"} Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.185750 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-1" podStartSLOduration=-9223371935.669054 podStartE2EDuration="1m41.185721553s" podCreationTimestamp="2026-03-17 15:59:46 +0000 UTC" firstStartedPulling="2026-03-17 15:59:51.797516088 +0000 UTC m=+1383.210832135" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:01:27.182295231 +0000 UTC m=+1478.595611288" watchObservedRunningTime="2026-03-17 16:01:27.185721553 +0000 UTC m=+1478.599037600" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.441373 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-xt87s"] Mar 17 16:01:27 crc kubenswrapper[4767]: E0317 16:01:27.441986 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f745e6a5-fa5d-4bee-9df5-df119632fbf0" containerName="mariadb-database-create" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.442025 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f745e6a5-fa5d-4bee-9df5-df119632fbf0" containerName="mariadb-database-create" Mar 17 16:01:27 crc kubenswrapper[4767]: E0317 16:01:27.442047 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae4f3fba-7231-451f-a3cd-d560c85212dd" containerName="mariadb-account-create-update" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.442072 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae4f3fba-7231-451f-a3cd-d560c85212dd" containerName="mariadb-account-create-update" Mar 17 16:01:27 crc kubenswrapper[4767]: E0317 16:01:27.442084 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ed389a-fa13-445f-826e-a88c04ef059d" containerName="mariadb-database-create" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.442091 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ed389a-fa13-445f-826e-a88c04ef059d" containerName="mariadb-database-create" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.442333 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="33ed389a-fa13-445f-826e-a88c04ef059d" containerName="mariadb-database-create" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.442360 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f745e6a5-fa5d-4bee-9df5-df119632fbf0" containerName="mariadb-database-create" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.442369 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae4f3fba-7231-451f-a3cd-d560c85212dd" containerName="mariadb-account-create-update" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.443196 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xt87s" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.458873 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.489056 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-xt87s"] Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.525016 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff7518ae-245d-419e-9328-58420b43d6a7-operator-scripts\") pod \"root-account-create-update-xt87s\" (UID: \"ff7518ae-245d-419e-9328-58420b43d6a7\") " pod="openstack/root-account-create-update-xt87s" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.525331 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsdwc\" (UniqueName: \"kubernetes.io/projected/ff7518ae-245d-419e-9328-58420b43d6a7-kube-api-access-vsdwc\") pod \"root-account-create-update-xt87s\" (UID: \"ff7518ae-245d-419e-9328-58420b43d6a7\") " pod="openstack/root-account-create-update-xt87s" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.553418 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.627606 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff7518ae-245d-419e-9328-58420b43d6a7-operator-scripts\") pod \"root-account-create-update-xt87s\" (UID: \"ff7518ae-245d-419e-9328-58420b43d6a7\") " pod="openstack/root-account-create-update-xt87s" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.627927 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsdwc\" (UniqueName: \"kubernetes.io/projected/ff7518ae-245d-419e-9328-58420b43d6a7-kube-api-access-vsdwc\") pod \"root-account-create-update-xt87s\" (UID: \"ff7518ae-245d-419e-9328-58420b43d6a7\") " pod="openstack/root-account-create-update-xt87s" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.629153 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff7518ae-245d-419e-9328-58420b43d6a7-operator-scripts\") pod \"root-account-create-update-xt87s\" (UID: \"ff7518ae-245d-419e-9328-58420b43d6a7\") " pod="openstack/root-account-create-update-xt87s" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.726819 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsdwc\" (UniqueName: \"kubernetes.io/projected/ff7518ae-245d-419e-9328-58420b43d6a7-kube-api-access-vsdwc\") pod \"root-account-create-update-xt87s\" (UID: \"ff7518ae-245d-419e-9328-58420b43d6a7\") " pod="openstack/root-account-create-update-xt87s" Mar 17 16:01:27 crc kubenswrapper[4767]: I0317 16:01:27.766216 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xt87s" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.188873 4767 generic.go:334] "Generic (PLEG): container finished" podID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerID="a150a9a169d4d790c0367da0e3743c173a7add3e0702539d16e0be602f1d139a" exitCode=0 Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.189554 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a082b2bf-8b94-40e7-be0c-be64f75a4c3e","Type":"ContainerDied","Data":"a150a9a169d4d790c0367da0e3743c173a7add3e0702539d16e0be602f1d139a"} Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.507608 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-xt87s"] Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.704516 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.772190 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89f09b66-7eec-4ad8-9116-83fa9d80fead-operator-scripts\") pod \"89f09b66-7eec-4ad8-9116-83fa9d80fead\" (UID: \"89f09b66-7eec-4ad8-9116-83fa9d80fead\") " Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.772308 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ctns\" (UniqueName: \"kubernetes.io/projected/89f09b66-7eec-4ad8-9116-83fa9d80fead-kube-api-access-8ctns\") pod \"89f09b66-7eec-4ad8-9116-83fa9d80fead\" (UID: \"89f09b66-7eec-4ad8-9116-83fa9d80fead\") " Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.777645 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89f09b66-7eec-4ad8-9116-83fa9d80fead-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "89f09b66-7eec-4ad8-9116-83fa9d80fead" (UID: "89f09b66-7eec-4ad8-9116-83fa9d80fead"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.780477 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89f09b66-7eec-4ad8-9116-83fa9d80fead-kube-api-access-8ctns" (OuterVolumeSpecName: "kube-api-access-8ctns") pod "89f09b66-7eec-4ad8-9116-83fa9d80fead" (UID: "89f09b66-7eec-4ad8-9116-83fa9d80fead"). InnerVolumeSpecName "kube-api-access-8ctns". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.876314 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89f09b66-7eec-4ad8-9116-83fa9d80fead-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.876359 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ctns\" (UniqueName: \"kubernetes.io/projected/89f09b66-7eec-4ad8-9116-83fa9d80fead-kube-api-access-8ctns\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.921234 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4hdjg"] Mar 17 16:01:28 crc kubenswrapper[4767]: E0317 16:01:28.921913 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89f09b66-7eec-4ad8-9116-83fa9d80fead" containerName="mariadb-account-create-update" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.921944 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="89f09b66-7eec-4ad8-9116-83fa9d80fead" containerName="mariadb-account-create-update" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.922277 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="89f09b66-7eec-4ad8-9116-83fa9d80fead" containerName="mariadb-account-create-update" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.923494 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.932508 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-hp85d" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.933414 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 17 16:01:28 crc kubenswrapper[4767]: I0317 16:01:28.946777 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4hdjg"] Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.083203 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-db-sync-config-data\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.083266 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r75dp\" (UniqueName: \"kubernetes.io/projected/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-kube-api-access-r75dp\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.083546 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-config-data\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.083981 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-combined-ca-bundle\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.187222 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r75dp\" (UniqueName: \"kubernetes.io/projected/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-kube-api-access-r75dp\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.187673 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-config-data\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.187815 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-combined-ca-bundle\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.187914 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-db-sync-config-data\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.193467 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-config-data\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.193567 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-combined-ca-bundle\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.218830 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-db-sync-config-data\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.257625 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-xt87s" event={"ID":"ff7518ae-245d-419e-9328-58420b43d6a7","Type":"ContainerStarted","Data":"91e2edd07554b3b5796895b4011d3d7141c61280aa1ac94d4e74b222b19dd295"} Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.257686 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-xt87s" event={"ID":"ff7518ae-245d-419e-9328-58420b43d6a7","Type":"ContainerStarted","Data":"bb8f440c133a2a4a19f61719cbc4629a53502c38b0974ee27f182558d07ebd21"} Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.276695 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a082b2bf-8b94-40e7-be0c-be64f75a4c3e","Type":"ContainerStarted","Data":"691fea73d6a12e81a25a85a1e6ada57dc114881e572a3370e586226a5dec638f"} Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.277264 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-2" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.312718 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" event={"ID":"89f09b66-7eec-4ad8-9116-83fa9d80fead","Type":"ContainerDied","Data":"a73da47cfb9e20a900b769ea12ce1982e22a749525498dac7263ef08043273b6"} Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.312779 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a73da47cfb9e20a900b769ea12ce1982e22a749525498dac7263ef08043273b6" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.312853 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-fa1b-account-create-update-nkxnj" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.315017 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r75dp\" (UniqueName: \"kubernetes.io/projected/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-kube-api-access-r75dp\") pod \"glance-db-sync-4hdjg\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.327057 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-xt87s" podStartSLOduration=2.3270277520000002 podStartE2EDuration="2.327027752s" podCreationTimestamp="2026-03-17 16:01:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:01:29.310819739 +0000 UTC m=+1480.724135796" watchObservedRunningTime="2026-03-17 16:01:29.327027752 +0000 UTC m=+1480.740343799" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.344919 4767 generic.go:334] "Generic (PLEG): container finished" podID="500b62ad-a562-419e-a44a-85030e12bf8b" containerID="c04d1896324154528cb7fccb824fd1e8af7ff5b8cab07299774c184cf2bd0702" exitCode=0 Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.344961 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-cqcdm" event={"ID":"500b62ad-a562-419e-a44a-85030e12bf8b","Type":"ContainerDied","Data":"c04d1896324154528cb7fccb824fd1e8af7ff5b8cab07299774c184cf2bd0702"} Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.386878 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.388314 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e2d93f34-029a-4de9-941f-e04a7c957fc3","Type":"ContainerStarted","Data":"db0e159f63364fd6fa158bd969553e8dee5b4565d6cd13db59f36b167fe2b73f"} Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.388368 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e2d93f34-029a-4de9-941f-e04a7c957fc3","Type":"ContainerStarted","Data":"4d9214bd9aa403560d28ac0049a89f83524355e4d4e2dbb5b9c20d44c80fedd5"} Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.388391 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.409529 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e718a8e4-5847-4051-a7d3-d01f0c22d801-operator-scripts\") pod \"e718a8e4-5847-4051-a7d3-d01f0c22d801\" (UID: \"e718a8e4-5847-4051-a7d3-d01f0c22d801\") " Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.409741 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghdbc\" (UniqueName: \"kubernetes.io/projected/e718a8e4-5847-4051-a7d3-d01f0c22d801-kube-api-access-ghdbc\") pod \"e718a8e4-5847-4051-a7d3-d01f0c22d801\" (UID: \"e718a8e4-5847-4051-a7d3-d01f0c22d801\") " Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.410265 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e718a8e4-5847-4051-a7d3-d01f0c22d801-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e718a8e4-5847-4051-a7d3-d01f0c22d801" (UID: "e718a8e4-5847-4051-a7d3-d01f0c22d801"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.410685 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e718a8e4-5847-4051-a7d3-d01f0c22d801-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.422088 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e718a8e4-5847-4051-a7d3-d01f0c22d801-kube-api-access-ghdbc" (OuterVolumeSpecName: "kube-api-access-ghdbc") pod "e718a8e4-5847-4051-a7d3-d01f0c22d801" (UID: "e718a8e4-5847-4051-a7d3-d01f0c22d801"). InnerVolumeSpecName "kube-api-access-ghdbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.423426 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-2" podStartSLOduration=-9223371933.431374 podStartE2EDuration="1m43.423401685s" podCreationTimestamp="2026-03-17 15:59:46 +0000 UTC" firstStartedPulling="2026-03-17 15:59:53.170901977 +0000 UTC m=+1384.584218024" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:01:29.420912168 +0000 UTC m=+1480.834228225" watchObservedRunningTime="2026-03-17 16:01:29.423401685 +0000 UTC m=+1480.836717732" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.512962 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghdbc\" (UniqueName: \"kubernetes.io/projected/e718a8e4-5847-4051-a7d3-d01f0c22d801-kube-api-access-ghdbc\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.559933 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4hdjg" Mar 17 16:01:29 crc kubenswrapper[4767]: I0317 16:01:29.592919 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=8.343292958 podStartE2EDuration="9.59289283s" podCreationTimestamp="2026-03-17 16:01:20 +0000 UTC" firstStartedPulling="2026-03-17 16:01:26.495483424 +0000 UTC m=+1477.908799481" lastFinishedPulling="2026-03-17 16:01:27.745083306 +0000 UTC m=+1479.158399353" observedRunningTime="2026-03-17 16:01:29.558737288 +0000 UTC m=+1480.972053345" watchObservedRunningTime="2026-03-17 16:01:29.59289283 +0000 UTC m=+1481.006208877" Mar 17 16:01:30 crc kubenswrapper[4767]: I0317 16:01:30.389664 4767 generic.go:334] "Generic (PLEG): container finished" podID="ff7518ae-245d-419e-9328-58420b43d6a7" containerID="91e2edd07554b3b5796895b4011d3d7141c61280aa1ac94d4e74b222b19dd295" exitCode=0 Mar 17 16:01:30 crc kubenswrapper[4767]: I0317 16:01:30.390058 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-xt87s" event={"ID":"ff7518ae-245d-419e-9328-58420b43d6a7","Type":"ContainerDied","Data":"91e2edd07554b3b5796895b4011d3d7141c61280aa1ac94d4e74b222b19dd295"} Mar 17 16:01:30 crc kubenswrapper[4767]: I0317 16:01:30.393954 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" Mar 17 16:01:30 crc kubenswrapper[4767]: I0317 16:01:30.402720 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f" event={"ID":"e718a8e4-5847-4051-a7d3-d01f0c22d801","Type":"ContainerDied","Data":"de89fcdeac5695b7e0d912533d08bed9f10405695969442d878ce1406ed3427a"} Mar 17 16:01:30 crc kubenswrapper[4767]: I0317 16:01:30.403007 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de89fcdeac5695b7e0d912533d08bed9f10405695969442d878ce1406ed3427a" Mar 17 16:01:30 crc kubenswrapper[4767]: I0317 16:01:30.574113 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4hdjg"] Mar 17 16:01:30 crc kubenswrapper[4767]: I0317 16:01:30.933617 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.088756 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/500b62ad-a562-419e-a44a-85030e12bf8b-etc-swift\") pod \"500b62ad-a562-419e-a44a-85030e12bf8b\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.088967 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-dispersionconf\") pod \"500b62ad-a562-419e-a44a-85030e12bf8b\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.088994 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6znwv\" (UniqueName: \"kubernetes.io/projected/500b62ad-a562-419e-a44a-85030e12bf8b-kube-api-access-6znwv\") pod \"500b62ad-a562-419e-a44a-85030e12bf8b\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.089037 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-combined-ca-bundle\") pod \"500b62ad-a562-419e-a44a-85030e12bf8b\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.089061 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-swiftconf\") pod \"500b62ad-a562-419e-a44a-85030e12bf8b\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.089133 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-scripts\") pod \"500b62ad-a562-419e-a44a-85030e12bf8b\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.089156 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-ring-data-devices\") pod \"500b62ad-a562-419e-a44a-85030e12bf8b\" (UID: \"500b62ad-a562-419e-a44a-85030e12bf8b\") " Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.089792 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/500b62ad-a562-419e-a44a-85030e12bf8b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "500b62ad-a562-419e-a44a-85030e12bf8b" (UID: "500b62ad-a562-419e-a44a-85030e12bf8b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.089956 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "500b62ad-a562-419e-a44a-85030e12bf8b" (UID: "500b62ad-a562-419e-a44a-85030e12bf8b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.125216 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/500b62ad-a562-419e-a44a-85030e12bf8b-kube-api-access-6znwv" (OuterVolumeSpecName: "kube-api-access-6znwv") pod "500b62ad-a562-419e-a44a-85030e12bf8b" (UID: "500b62ad-a562-419e-a44a-85030e12bf8b"). InnerVolumeSpecName "kube-api-access-6znwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.129639 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "500b62ad-a562-419e-a44a-85030e12bf8b" (UID: "500b62ad-a562-419e-a44a-85030e12bf8b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.130454 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "500b62ad-a562-419e-a44a-85030e12bf8b" (UID: "500b62ad-a562-419e-a44a-85030e12bf8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.132418 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "500b62ad-a562-419e-a44a-85030e12bf8b" (UID: "500b62ad-a562-419e-a44a-85030e12bf8b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.143914 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-scripts" (OuterVolumeSpecName: "scripts") pod "500b62ad-a562-419e-a44a-85030e12bf8b" (UID: "500b62ad-a562-419e-a44a-85030e12bf8b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.192300 4767 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.192335 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.192348 4767 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/500b62ad-a562-419e-a44a-85030e12bf8b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.192368 4767 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/500b62ad-a562-419e-a44a-85030e12bf8b-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.192381 4767 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.192391 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6znwv\" (UniqueName: \"kubernetes.io/projected/500b62ad-a562-419e-a44a-85030e12bf8b-kube-api-access-6znwv\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.192400 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500b62ad-a562-419e-a44a-85030e12bf8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.230504 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.241592 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-kqr5p" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.405998 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-cqcdm" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.405997 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-cqcdm" event={"ID":"500b62ad-a562-419e-a44a-85030e12bf8b","Type":"ContainerDied","Data":"db38d47c5fda62600a620d5a2a323a5cd2e83faf6ff7061d977148ec4cbfb4a2"} Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.406072 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db38d47c5fda62600a620d5a2a323a5cd2e83faf6ff7061d977148ec4cbfb4a2" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.410841 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4hdjg" event={"ID":"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa","Type":"ContainerStarted","Data":"19f74ed0ab1b263431de35de8ce8745096f4bb14acb0ecc174687ae0c8bb7dd5"} Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.481359 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-47tcm" podUID="d0bf7fbd-b9f4-4b6d-8d2e-f3e299f71cc0" containerName="ovn-controller" probeResult="failure" output=< Mar 17 16:01:31 crc kubenswrapper[4767]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 17 16:01:31 crc kubenswrapper[4767]: > Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.509273 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-47tcm-config-fkqzz"] Mar 17 16:01:31 crc kubenswrapper[4767]: E0317 16:01:31.509824 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="500b62ad-a562-419e-a44a-85030e12bf8b" containerName="swift-ring-rebalance" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.509844 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="500b62ad-a562-419e-a44a-85030e12bf8b" containerName="swift-ring-rebalance" Mar 17 16:01:31 crc kubenswrapper[4767]: E0317 16:01:31.509864 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e718a8e4-5847-4051-a7d3-d01f0c22d801" containerName="mariadb-database-create" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.509871 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e718a8e4-5847-4051-a7d3-d01f0c22d801" containerName="mariadb-database-create" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.510140 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="e718a8e4-5847-4051-a7d3-d01f0c22d801" containerName="mariadb-database-create" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.510153 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="500b62ad-a562-419e-a44a-85030e12bf8b" containerName="swift-ring-rebalance" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.511077 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.517252 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.560278 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-47tcm-config-fkqzz"] Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.602497 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run-ovn\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.602832 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-988lp\" (UniqueName: \"kubernetes.io/projected/47b5d4a3-d4f6-4fb5-a166-442042e92045-kube-api-access-988lp\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.602968 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-scripts\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.603052 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-log-ovn\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.603082 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.603102 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-additional-scripts\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.705534 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-988lp\" (UniqueName: \"kubernetes.io/projected/47b5d4a3-d4f6-4fb5-a166-442042e92045-kube-api-access-988lp\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.705645 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-scripts\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.705721 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-log-ovn\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.705745 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.705764 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-additional-scripts\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.705790 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run-ovn\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.706163 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run-ovn\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.707014 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.707088 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-log-ovn\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.707669 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-additional-scripts\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.708770 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-scripts\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.738150 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-988lp\" (UniqueName: \"kubernetes.io/projected/47b5d4a3-d4f6-4fb5-a166-442042e92045-kube-api-access-988lp\") pod \"ovn-controller-47tcm-config-fkqzz\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.839588 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:31 crc kubenswrapper[4767]: I0317 16:01:31.988089 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xt87s" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.115475 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff7518ae-245d-419e-9328-58420b43d6a7-operator-scripts\") pod \"ff7518ae-245d-419e-9328-58420b43d6a7\" (UID: \"ff7518ae-245d-419e-9328-58420b43d6a7\") " Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.116082 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsdwc\" (UniqueName: \"kubernetes.io/projected/ff7518ae-245d-419e-9328-58420b43d6a7-kube-api-access-vsdwc\") pod \"ff7518ae-245d-419e-9328-58420b43d6a7\" (UID: \"ff7518ae-245d-419e-9328-58420b43d6a7\") " Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.116286 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7518ae-245d-419e-9328-58420b43d6a7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ff7518ae-245d-419e-9328-58420b43d6a7" (UID: "ff7518ae-245d-419e-9328-58420b43d6a7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.117122 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff7518ae-245d-419e-9328-58420b43d6a7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.141922 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff7518ae-245d-419e-9328-58420b43d6a7-kube-api-access-vsdwc" (OuterVolumeSpecName: "kube-api-access-vsdwc") pod "ff7518ae-245d-419e-9328-58420b43d6a7" (UID: "ff7518ae-245d-419e-9328-58420b43d6a7"). InnerVolumeSpecName "kube-api-access-vsdwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.219719 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsdwc\" (UniqueName: \"kubernetes.io/projected/ff7518ae-245d-419e-9328-58420b43d6a7-kube-api-access-vsdwc\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.431532 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-xt87s" event={"ID":"ff7518ae-245d-419e-9328-58420b43d6a7","Type":"ContainerDied","Data":"bb8f440c133a2a4a19f61719cbc4629a53502c38b0974ee27f182558d07ebd21"} Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.431590 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb8f440c133a2a4a19f61719cbc4629a53502c38b0974ee27f182558d07ebd21" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.431677 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xt87s" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.494072 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Mar 17 16:01:32 crc kubenswrapper[4767]: E0317 16:01:32.494670 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7518ae-245d-419e-9328-58420b43d6a7" containerName="mariadb-account-create-update" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.494692 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7518ae-245d-419e-9328-58420b43d6a7" containerName="mariadb-account-create-update" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.494900 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff7518ae-245d-419e-9328-58420b43d6a7" containerName="mariadb-account-create-update" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.495971 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.505925 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.555863 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.618136 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.618665 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-config-data\") pod \"mysqld-exporter-0\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " pod="openstack/mysqld-exporter-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.626088 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.626920 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dr5v\" (UniqueName: \"kubernetes.io/projected/bf461aeb-5692-40f0-be55-16e509064dc6-kube-api-access-4dr5v\") pod \"mysqld-exporter-0\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " pod="openstack/mysqld-exporter-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.628506 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " pod="openstack/mysqld-exporter-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.628999 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-47tcm-config-fkqzz"] Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.660668 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.737596 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dr5v\" (UniqueName: \"kubernetes.io/projected/bf461aeb-5692-40f0-be55-16e509064dc6-kube-api-access-4dr5v\") pod \"mysqld-exporter-0\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " pod="openstack/mysqld-exporter-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.737838 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " pod="openstack/mysqld-exporter-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.738154 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-config-data\") pod \"mysqld-exporter-0\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " pod="openstack/mysqld-exporter-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.764721 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " pod="openstack/mysqld-exporter-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.767544 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-config-data\") pod \"mysqld-exporter-0\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " pod="openstack/mysqld-exporter-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.768127 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dr5v\" (UniqueName: \"kubernetes.io/projected/bf461aeb-5692-40f0-be55-16e509064dc6-kube-api-access-4dr5v\") pod \"mysqld-exporter-0\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " pod="openstack/mysqld-exporter-0" Mar 17 16:01:32 crc kubenswrapper[4767]: I0317 16:01:32.981620 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 17 16:01:33 crc kubenswrapper[4767]: I0317 16:01:33.483775 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-47tcm-config-fkqzz" event={"ID":"47b5d4a3-d4f6-4fb5-a166-442042e92045","Type":"ContainerStarted","Data":"51f03ebbb40543f93b81461057db4f82070357de7e7ccc7088777523c2cc5c8b"} Mar 17 16:01:33 crc kubenswrapper[4767]: I0317 16:01:33.483904 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-47tcm-config-fkqzz" event={"ID":"47b5d4a3-d4f6-4fb5-a166-442042e92045","Type":"ContainerStarted","Data":"79f8991c6df4d76d8b2a8c5bd0fa0eba1e81bcc1ab733ec375d96c74333b0ae7"} Mar 17 16:01:33 crc kubenswrapper[4767]: I0317 16:01:33.784255 4767 scope.go:117] "RemoveContainer" containerID="7c3acd80733a58dbc10ddc2eea984975cf074e58e075379c6fb58bf80bc50395" Mar 17 16:01:33 crc kubenswrapper[4767]: I0317 16:01:33.873901 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-47tcm-config-fkqzz" podStartSLOduration=2.873867126 podStartE2EDuration="2.873867126s" podCreationTimestamp="2026-03-17 16:01:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:01:33.548210001 +0000 UTC m=+1484.961526048" watchObservedRunningTime="2026-03-17 16:01:33.873867126 +0000 UTC m=+1485.287183173" Mar 17 16:01:33 crc kubenswrapper[4767]: I0317 16:01:33.906741 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 17 16:01:34 crc kubenswrapper[4767]: I0317 16:01:34.496670 4767 generic.go:334] "Generic (PLEG): container finished" podID="47b5d4a3-d4f6-4fb5-a166-442042e92045" containerID="51f03ebbb40543f93b81461057db4f82070357de7e7ccc7088777523c2cc5c8b" exitCode=0 Mar 17 16:01:34 crc kubenswrapper[4767]: I0317 16:01:34.496905 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-47tcm-config-fkqzz" event={"ID":"47b5d4a3-d4f6-4fb5-a166-442042e92045","Type":"ContainerDied","Data":"51f03ebbb40543f93b81461057db4f82070357de7e7ccc7088777523c2cc5c8b"} Mar 17 16:01:34 crc kubenswrapper[4767]: I0317 16:01:34.501283 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"bf461aeb-5692-40f0-be55-16e509064dc6","Type":"ContainerStarted","Data":"386ce5f130b75578e149d4e9df0c9e87531af50c61593416f9699a7ea301ef8b"} Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.249409 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.303041 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-scripts\") pod \"47b5d4a3-d4f6-4fb5-a166-442042e92045\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.303256 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run-ovn\") pod \"47b5d4a3-d4f6-4fb5-a166-442042e92045\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.303307 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-additional-scripts\") pod \"47b5d4a3-d4f6-4fb5-a166-442042e92045\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.303332 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-log-ovn\") pod \"47b5d4a3-d4f6-4fb5-a166-442042e92045\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.303378 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run\") pod \"47b5d4a3-d4f6-4fb5-a166-442042e92045\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.303368 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "47b5d4a3-d4f6-4fb5-a166-442042e92045" (UID: "47b5d4a3-d4f6-4fb5-a166-442042e92045"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.303468 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "47b5d4a3-d4f6-4fb5-a166-442042e92045" (UID: "47b5d4a3-d4f6-4fb5-a166-442042e92045"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.303478 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-988lp\" (UniqueName: \"kubernetes.io/projected/47b5d4a3-d4f6-4fb5-a166-442042e92045-kube-api-access-988lp\") pod \"47b5d4a3-d4f6-4fb5-a166-442042e92045\" (UID: \"47b5d4a3-d4f6-4fb5-a166-442042e92045\") " Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.303494 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run" (OuterVolumeSpecName: "var-run") pod "47b5d4a3-d4f6-4fb5-a166-442042e92045" (UID: "47b5d4a3-d4f6-4fb5-a166-442042e92045"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.304113 4767 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.304128 4767 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.304138 4767 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47b5d4a3-d4f6-4fb5-a166-442042e92045-var-run\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.304127 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "47b5d4a3-d4f6-4fb5-a166-442042e92045" (UID: "47b5d4a3-d4f6-4fb5-a166-442042e92045"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.304478 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-scripts" (OuterVolumeSpecName: "scripts") pod "47b5d4a3-d4f6-4fb5-a166-442042e92045" (UID: "47b5d4a3-d4f6-4fb5-a166-442042e92045"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.312062 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47b5d4a3-d4f6-4fb5-a166-442042e92045-kube-api-access-988lp" (OuterVolumeSpecName: "kube-api-access-988lp") pod "47b5d4a3-d4f6-4fb5-a166-442042e92045" (UID: "47b5d4a3-d4f6-4fb5-a166-442042e92045"). InnerVolumeSpecName "kube-api-access-988lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.406744 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.406789 4767 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/47b5d4a3-d4f6-4fb5-a166-442042e92045-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.406806 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-988lp\" (UniqueName: \"kubernetes.io/projected/47b5d4a3-d4f6-4fb5-a166-442042e92045-kube-api-access-988lp\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.498513 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-47tcm" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.546480 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-47tcm-config-fkqzz" event={"ID":"47b5d4a3-d4f6-4fb5-a166-442042e92045","Type":"ContainerDied","Data":"79f8991c6df4d76d8b2a8c5bd0fa0eba1e81bcc1ab733ec375d96c74333b0ae7"} Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.546528 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-47tcm-config-fkqzz" Mar 17 16:01:36 crc kubenswrapper[4767]: I0317 16:01:36.546539 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79f8991c6df4d76d8b2a8c5bd0fa0eba1e81bcc1ab733ec375d96c74333b0ae7" Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.377687 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.378650 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="prometheus" containerID="cri-o://9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616" gracePeriod=600 Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.379156 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="thanos-sidecar" containerID="cri-o://f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05" gracePeriod=600 Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.379250 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="config-reloader" containerID="cri-o://66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de" gracePeriod=600 Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.461071 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-47tcm-config-fkqzz"] Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.474228 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-47tcm-config-fkqzz"] Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.554509 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.145:9090/-/ready\": dial tcp 10.217.0.145:9090: connect: connection refused" Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.571477 4767 generic.go:334] "Generic (PLEG): container finished" podID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerID="f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05" exitCode=0 Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.571524 4767 generic.go:334] "Generic (PLEG): container finished" podID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerID="9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616" exitCode=0 Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.571613 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f","Type":"ContainerDied","Data":"f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05"} Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.571654 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f","Type":"ContainerDied","Data":"9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616"} Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.581520 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"bf461aeb-5692-40f0-be55-16e509064dc6","Type":"ContainerStarted","Data":"1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8"} Mar 17 16:01:37 crc kubenswrapper[4767]: I0317 16:01:37.612038 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.597490126 podStartE2EDuration="5.612000308s" podCreationTimestamp="2026-03-17 16:01:32 +0000 UTC" firstStartedPulling="2026-03-17 16:01:33.910991587 +0000 UTC m=+1485.324307634" lastFinishedPulling="2026-03-17 16:01:36.925501769 +0000 UTC m=+1488.338817816" observedRunningTime="2026-03-17 16:01:37.605917295 +0000 UTC m=+1489.019233342" watchObservedRunningTime="2026-03-17 16:01:37.612000308 +0000 UTC m=+1489.025316355" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.518442 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.567260 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-thanos-prometheus-http-client-file\") pod \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.567450 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config-out\") pod \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.567656 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-tls-assets\") pod \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.567759 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-2\") pod \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.567906 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-1\") pod \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.568020 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config\") pod \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.568373 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\") pod \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.568448 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-0\") pod \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.568471 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sbfg\" (UniqueName: \"kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-kube-api-access-4sbfg\") pod \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.568512 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-web-config\") pod \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\" (UID: \"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f\") " Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.568718 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-2" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-2") pod "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" (UID: "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.569343 4767 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-2\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.570460 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-1" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-1") pod "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" (UID: "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.570726 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" (UID: "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.577953 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" (UID: "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.578744 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-kube-api-access-4sbfg" (OuterVolumeSpecName: "kube-api-access-4sbfg") pod "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" (UID: "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f"). InnerVolumeSpecName "kube-api-access-4sbfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.580613 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" (UID: "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.592411 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config" (OuterVolumeSpecName: "config") pod "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" (UID: "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.602955 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config-out" (OuterVolumeSpecName: "config-out") pod "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" (UID: "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.634759 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" (UID: "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f"). InnerVolumeSpecName "pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.645655 4767 generic.go:334] "Generic (PLEG): container finished" podID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerID="66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de" exitCode=0 Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.646302 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.646314 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f","Type":"ContainerDied","Data":"66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de"} Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.646769 4767 scope.go:117] "RemoveContainer" containerID="f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.646752 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f07d3d0-a5f9-49c0-adf4-f748cbe5947f","Type":"ContainerDied","Data":"6ff9ea01d5c579f3323ef089293e6ffdc95b914337844d915782fa2e02bdcdb3"} Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.673138 4767 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.673215 4767 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config-out\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.673243 4767 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-tls-assets\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.673267 4767 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-1\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.673295 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.673354 4767 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\") on node \"crc\" " Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.673372 4767 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.673392 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sbfg\" (UniqueName: \"kubernetes.io/projected/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-kube-api-access-4sbfg\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.674556 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-web-config" (OuterVolumeSpecName: "web-config") pod "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" (UID: "2f07d3d0-a5f9-49c0-adf4-f748cbe5947f"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.729022 4767 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.729212 4767 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62") on node "crc" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.760122 4767 scope.go:117] "RemoveContainer" containerID="66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.776352 4767 reconciler_common.go:293] "Volume detached for volume \"pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.776612 4767 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f-web-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.784888 4767 scope.go:117] "RemoveContainer" containerID="9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.817726 4767 scope.go:117] "RemoveContainer" containerID="5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.848643 4767 scope.go:117] "RemoveContainer" containerID="f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05" Mar 17 16:01:38 crc kubenswrapper[4767]: E0317 16:01:38.849931 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05\": container with ID starting with f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05 not found: ID does not exist" containerID="f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.850001 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05"} err="failed to get container status \"f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05\": rpc error: code = NotFound desc = could not find container \"f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05\": container with ID starting with f6b2c034346d2015407012f671935c7e6e81441aac1ced6ea2fa183ed7953b05 not found: ID does not exist" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.850041 4767 scope.go:117] "RemoveContainer" containerID="66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de" Mar 17 16:01:38 crc kubenswrapper[4767]: E0317 16:01:38.850863 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de\": container with ID starting with 66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de not found: ID does not exist" containerID="66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.850920 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de"} err="failed to get container status \"66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de\": rpc error: code = NotFound desc = could not find container \"66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de\": container with ID starting with 66eb4a636dc40455ffc12340092ec0be4efef6cedefaa9b0a72fbaba482101de not found: ID does not exist" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.850958 4767 scope.go:117] "RemoveContainer" containerID="9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616" Mar 17 16:01:38 crc kubenswrapper[4767]: E0317 16:01:38.852671 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616\": container with ID starting with 9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616 not found: ID does not exist" containerID="9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.852797 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616"} err="failed to get container status \"9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616\": rpc error: code = NotFound desc = could not find container \"9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616\": container with ID starting with 9bf503f63ff3ff98293476567c403075f627e65b98dfc7913e7716b9cdb7e616 not found: ID does not exist" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.852905 4767 scope.go:117] "RemoveContainer" containerID="5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0" Mar 17 16:01:38 crc kubenswrapper[4767]: E0317 16:01:38.853488 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0\": container with ID starting with 5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0 not found: ID does not exist" containerID="5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0" Mar 17 16:01:38 crc kubenswrapper[4767]: I0317 16:01:38.853554 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0"} err="failed to get container status \"5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0\": rpc error: code = NotFound desc = could not find container \"5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0\": container with ID starting with 5c2b8c4229fccf0f31a13d8cbf50c4cb1b54df889c0452398a444f2919cd58e0 not found: ID does not exist" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.006977 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.017696 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.048262 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 17 16:01:39 crc kubenswrapper[4767]: E0317 16:01:39.048833 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="config-reloader" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.048867 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="config-reloader" Mar 17 16:01:39 crc kubenswrapper[4767]: E0317 16:01:39.048876 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="prometheus" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.048882 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="prometheus" Mar 17 16:01:39 crc kubenswrapper[4767]: E0317 16:01:39.048912 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="thanos-sidecar" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.048920 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="thanos-sidecar" Mar 17 16:01:39 crc kubenswrapper[4767]: E0317 16:01:39.048945 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="init-config-reloader" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.048951 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="init-config-reloader" Mar 17 16:01:39 crc kubenswrapper[4767]: E0317 16:01:39.048967 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b5d4a3-d4f6-4fb5-a166-442042e92045" containerName="ovn-config" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.048972 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b5d4a3-d4f6-4fb5-a166-442042e92045" containerName="ovn-config" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.049207 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="thanos-sidecar" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.049223 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="47b5d4a3-d4f6-4fb5-a166-442042e92045" containerName="ovn-config" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.049236 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="config-reloader" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.049249 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" containerName="prometheus" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.051575 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.054391 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.054877 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.055209 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-97b2h" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.060839 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.061158 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.062917 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.063059 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.063218 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.064390 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082454 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/94b1e014-c27b-4b81-a12d-eecfe02d5b67-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082522 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082588 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/94b1e014-c27b-4b81-a12d-eecfe02d5b67-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082621 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-config\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082651 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/94b1e014-c27b-4b81-a12d-eecfe02d5b67-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082692 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082726 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zth96\" (UniqueName: \"kubernetes.io/projected/94b1e014-c27b-4b81-a12d-eecfe02d5b67-kube-api-access-zth96\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082806 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082838 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/94b1e014-c27b-4b81-a12d-eecfe02d5b67-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082862 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082916 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.082945 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/94b1e014-c27b-4b81-a12d-eecfe02d5b67-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.083025 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.092879 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185375 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185460 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/94b1e014-c27b-4b81-a12d-eecfe02d5b67-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185492 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185528 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/94b1e014-c27b-4b81-a12d-eecfe02d5b67-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185571 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-config\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185591 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/94b1e014-c27b-4b81-a12d-eecfe02d5b67-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185628 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185670 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zth96\" (UniqueName: \"kubernetes.io/projected/94b1e014-c27b-4b81-a12d-eecfe02d5b67-kube-api-access-zth96\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185747 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185781 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/94b1e014-c27b-4b81-a12d-eecfe02d5b67-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185800 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185820 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.185848 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/94b1e014-c27b-4b81-a12d-eecfe02d5b67-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.186736 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/94b1e014-c27b-4b81-a12d-eecfe02d5b67-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.187442 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/94b1e014-c27b-4b81-a12d-eecfe02d5b67-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.188218 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/94b1e014-c27b-4b81-a12d-eecfe02d5b67-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.191308 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-config\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.191359 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.191568 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.193807 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.193960 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/94b1e014-c27b-4b81-a12d-eecfe02d5b67-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.195467 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.195801 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94b1e014-c27b-4b81-a12d-eecfe02d5b67-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.197380 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.197421 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0b502474e72a7e0212a814002fbab995e6f4dc4694e1fb66ec323aa24c386d38/globalmount\"" pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.203156 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/94b1e014-c27b-4b81-a12d-eecfe02d5b67-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.210549 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zth96\" (UniqueName: \"kubernetes.io/projected/94b1e014-c27b-4b81-a12d-eecfe02d5b67-kube-api-access-zth96\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.245182 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-32b9bee2-05fe-4a0b-97e4-57ccbe086d62\") pod \"prometheus-metric-storage-0\" (UID: \"94b1e014-c27b-4b81-a12d-eecfe02d5b67\") " pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.298428 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="048fa918-95b2-404e-81f5-8989ed47ec0a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.132:5671: connect: connection refused" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.352408 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.133:5671: connect: connection refused" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.369326 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f07d3d0-a5f9-49c0-adf4-f748cbe5947f" path="/var/lib/kubelet/pods/2f07d3d0-a5f9-49c0-adf4-f748cbe5947f/volumes" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.370262 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47b5d4a3-d4f6-4fb5-a166-442042e92045" path="/var/lib/kubelet/pods/47b5d4a3-d4f6-4fb5-a166-442042e92045/volumes" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.376707 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:39 crc kubenswrapper[4767]: I0317 16:01:39.830349 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.134:5671: connect: connection refused" Mar 17 16:01:40 crc kubenswrapper[4767]: I0317 16:01:40.019208 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 17 16:01:40 crc kubenswrapper[4767]: W0317 16:01:40.034447 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94b1e014_c27b_4b81_a12d_eecfe02d5b67.slice/crio-f7b4aacc26ee33fbab083111182424e1926219529802cc9cb19ae720cd7ef5ff WatchSource:0}: Error finding container f7b4aacc26ee33fbab083111182424e1926219529802cc9cb19ae720cd7ef5ff: Status 404 returned error can't find the container with id f7b4aacc26ee33fbab083111182424e1926219529802cc9cb19ae720cd7ef5ff Mar 17 16:01:40 crc kubenswrapper[4767]: I0317 16:01:40.326543 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.135:5671: connect: connection refused" Mar 17 16:01:40 crc kubenswrapper[4767]: I0317 16:01:40.517540 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:40 crc kubenswrapper[4767]: I0317 16:01:40.524678 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/803604b0-bf06-4c1b-bdce-338284abc323-etc-swift\") pod \"swift-storage-0\" (UID: \"803604b0-bf06-4c1b-bdce-338284abc323\") " pod="openstack/swift-storage-0" Mar 17 16:01:40 crc kubenswrapper[4767]: I0317 16:01:40.666827 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 17 16:01:40 crc kubenswrapper[4767]: I0317 16:01:40.712745 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"94b1e014-c27b-4b81-a12d-eecfe02d5b67","Type":"ContainerStarted","Data":"f7b4aacc26ee33fbab083111182424e1926219529802cc9cb19ae720cd7ef5ff"} Mar 17 16:01:41 crc kubenswrapper[4767]: I0317 16:01:41.268520 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 17 16:01:41 crc kubenswrapper[4767]: W0317 16:01:41.281353 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod803604b0_bf06_4c1b_bdce_338284abc323.slice/crio-666d7584701051ee15c9798453b6e69f5f0306de708e30ff68eb8f88cd29a574 WatchSource:0}: Error finding container 666d7584701051ee15c9798453b6e69f5f0306de708e30ff68eb8f88cd29a574: Status 404 returned error can't find the container with id 666d7584701051ee15c9798453b6e69f5f0306de708e30ff68eb8f88cd29a574 Mar 17 16:01:41 crc kubenswrapper[4767]: I0317 16:01:41.462075 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 17 16:01:41 crc kubenswrapper[4767]: I0317 16:01:41.727733 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"666d7584701051ee15c9798453b6e69f5f0306de708e30ff68eb8f88cd29a574"} Mar 17 16:01:43 crc kubenswrapper[4767]: I0317 16:01:43.753114 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"94b1e014-c27b-4b81-a12d-eecfe02d5b67","Type":"ContainerStarted","Data":"ef5d8322af5317732fa70db44516c18520c80d6f12897c1cef50cef3f5d8876b"} Mar 17 16:01:49 crc kubenswrapper[4767]: I0317 16:01:49.297808 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="048fa918-95b2-404e-81f5-8989ed47ec0a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.132:5671: connect: connection refused" Mar 17 16:01:49 crc kubenswrapper[4767]: I0317 16:01:49.349207 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.133:5671: connect: connection refused" Mar 17 16:01:49 crc kubenswrapper[4767]: I0317 16:01:49.828493 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"94b1e014-c27b-4b81-a12d-eecfe02d5b67","Type":"ContainerDied","Data":"ef5d8322af5317732fa70db44516c18520c80d6f12897c1cef50cef3f5d8876b"} Mar 17 16:01:49 crc kubenswrapper[4767]: I0317 16:01:49.828436 4767 generic.go:334] "Generic (PLEG): container finished" podID="94b1e014-c27b-4b81-a12d-eecfe02d5b67" containerID="ef5d8322af5317732fa70db44516c18520c80d6f12897c1cef50cef3f5d8876b" exitCode=0 Mar 17 16:01:49 crc kubenswrapper[4767]: I0317 16:01:49.830543 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.134:5671: connect: connection refused" Mar 17 16:01:50 crc kubenswrapper[4767]: I0317 16:01:50.328215 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:01:50 crc kubenswrapper[4767]: I0317 16:01:50.844565 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"94b1e014-c27b-4b81-a12d-eecfe02d5b67","Type":"ContainerStarted","Data":"14d6418137a30155b33d7cedd95459d22cd9386c43f29132078e29822a4dcfcf"} Mar 17 16:01:50 crc kubenswrapper[4767]: I0317 16:01:50.846981 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4hdjg" event={"ID":"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa","Type":"ContainerStarted","Data":"e2ddeff074fda785b4914c153886c6fe3788507e73308757b73327834debf484"} Mar 17 16:01:50 crc kubenswrapper[4767]: I0317 16:01:50.849617 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"2282170db978e981eba6fe8c2b614e0a3c8844bfd63f894e773912d1d3d9dd63"} Mar 17 16:01:50 crc kubenswrapper[4767]: I0317 16:01:50.849665 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"7fef43bc3ac02dc86f64b0387ee7e2b5a9172fe735302962060fb5c0572e23b7"} Mar 17 16:01:50 crc kubenswrapper[4767]: I0317 16:01:50.877485 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4hdjg" podStartSLOduration=4.04209141 podStartE2EDuration="22.877452684s" podCreationTimestamp="2026-03-17 16:01:28 +0000 UTC" firstStartedPulling="2026-03-17 16:01:30.60418946 +0000 UTC m=+1482.017505507" lastFinishedPulling="2026-03-17 16:01:49.439550734 +0000 UTC m=+1500.852866781" observedRunningTime="2026-03-17 16:01:50.866144302 +0000 UTC m=+1502.279460379" watchObservedRunningTime="2026-03-17 16:01:50.877452684 +0000 UTC m=+1502.290768751" Mar 17 16:01:51 crc kubenswrapper[4767]: I0317 16:01:51.870775 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"d0046f06829ae9a7c2fbfe721140843a1324d063771fb36edf55ac3f00eef1e6"} Mar 17 16:01:51 crc kubenswrapper[4767]: I0317 16:01:51.872210 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"8e00b50a4f7f9105f387f3b2fbe78ba119af5e438b9f93389c6e570fed130354"} Mar 17 16:01:52 crc kubenswrapper[4767]: I0317 16:01:52.904043 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"df8693c2ba9838c3a812beef5e394b20ae24b1537ba2877ce5d2c2a46c6c1f8f"} Mar 17 16:01:53 crc kubenswrapper[4767]: I0317 16:01:53.919312 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"ea80addd7168c1b9551a4485ca000932a09a32ca26c209ee85ac84df08da414f"} Mar 17 16:01:53 crc kubenswrapper[4767]: I0317 16:01:53.919922 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"9ad09a4c37e5c78b8b7ebff9d4fa4782de163f9c3bae4f28730c163851dabb97"} Mar 17 16:01:53 crc kubenswrapper[4767]: I0317 16:01:53.921622 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"94b1e014-c27b-4b81-a12d-eecfe02d5b67","Type":"ContainerStarted","Data":"2969942da3a1ca18a85e75b0d2eb441e40ee162b2806276e868fefa2d2e19a7b"} Mar 17 16:01:54 crc kubenswrapper[4767]: I0317 16:01:54.936532 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"94b1e014-c27b-4b81-a12d-eecfe02d5b67","Type":"ContainerStarted","Data":"df390d1504eaa2e8954ba5280484fe2bcb12f2c6c8cfa994e6a0c3bfa2ca90ba"} Mar 17 16:01:54 crc kubenswrapper[4767]: I0317 16:01:54.968910 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"60ff41c1135b44b30b2b9b8c56de8ff7c2698f5eb8f717d702a460344dd87bb4"} Mar 17 16:01:55 crc kubenswrapper[4767]: I0317 16:01:55.005292 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.005263981 podStartE2EDuration="16.005263981s" podCreationTimestamp="2026-03-17 16:01:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:01:55.002190109 +0000 UTC m=+1506.415506176" watchObservedRunningTime="2026-03-17 16:01:55.005263981 +0000 UTC m=+1506.418580038" Mar 17 16:01:56 crc kubenswrapper[4767]: I0317 16:01:56.011345 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"86bf56ceba744dd27de492ce2d76c4ee512b6aae7ec0daecb4950c083631526f"} Mar 17 16:01:56 crc kubenswrapper[4767]: I0317 16:01:56.012035 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"2f498bfec592083f885bc950e9bc88891b34a523f73daa4ec62326c8fefd43e9"} Mar 17 16:01:56 crc kubenswrapper[4767]: I0317 16:01:56.012055 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"8ddc165eea0620fd67136840290d19e51bdd32ad167fe0fef82ae1763bf82aaa"} Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.032731 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"9e6b20de4282c0809486f149949d38baff018705510133bd335818ee5345693b"} Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.033323 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"4ba6e3c0f4b46ed9b647df3c308446c535b0e381af77763e892538419c676153"} Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.033337 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"cd4ae85ca000e6da0b3304a9ee4210f2b33b97e6bff9a67699f85c60571cebb4"} Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.033347 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"803604b0-bf06-4c1b-bdce-338284abc323","Type":"ContainerStarted","Data":"4cb6f6893d0fb8a22bc6122871729fc1d4f1b7c40a916dc022dbdc5555628887"} Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.077601 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.078311409 podStartE2EDuration="50.077579208s" podCreationTimestamp="2026-03-17 16:01:07 +0000 UTC" firstStartedPulling="2026-03-17 16:01:41.283535732 +0000 UTC m=+1492.696851789" lastFinishedPulling="2026-03-17 16:01:55.282803541 +0000 UTC m=+1506.696119588" observedRunningTime="2026-03-17 16:01:57.070054267 +0000 UTC m=+1508.483370324" watchObservedRunningTime="2026-03-17 16:01:57.077579208 +0000 UTC m=+1508.490895255" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.457294 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-nt244"] Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.459674 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.462377 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.486663 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-nt244"] Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.489878 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.489930 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.490014 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.490066 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-svc\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.490084 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98kkq\" (UniqueName: \"kubernetes.io/projected/e33e642c-58c5-4783-955a-aba880206b8c-kube-api-access-98kkq\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.490194 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-config\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.592444 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-svc\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.592515 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98kkq\" (UniqueName: \"kubernetes.io/projected/e33e642c-58c5-4783-955a-aba880206b8c-kube-api-access-98kkq\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.592614 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-config\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.592719 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.592736 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.592807 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.593716 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-svc\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.593803 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-config\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.594154 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.594159 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.594196 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.613699 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98kkq\" (UniqueName: \"kubernetes.io/projected/e33e642c-58c5-4783-955a-aba880206b8c-kube-api-access-98kkq\") pod \"dnsmasq-dns-764c5664d7-nt244\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:57 crc kubenswrapper[4767]: I0317 16:01:57.785605 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:01:58 crc kubenswrapper[4767]: I0317 16:01:58.361019 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-nt244"] Mar 17 16:01:58 crc kubenswrapper[4767]: W0317 16:01:58.361861 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode33e642c_58c5_4783_955a_aba880206b8c.slice/crio-9d77e033908c76917742719602639a8404808342751d16050d29a17fe2930f2b WatchSource:0}: Error finding container 9d77e033908c76917742719602639a8404808342751d16050d29a17fe2930f2b: Status 404 returned error can't find the container with id 9d77e033908c76917742719602639a8404808342751d16050d29a17fe2930f2b Mar 17 16:01:59 crc kubenswrapper[4767]: E0317 16:01:59.030706 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode33e642c_58c5_4783_955a_aba880206b8c.slice/crio-conmon-aeba7e474edd62db80c2ceb1ada7b1450076b2cd0c459187eac2aeca88bd8ca2.scope\": RecentStats: unable to find data in memory cache]" Mar 17 16:01:59 crc kubenswrapper[4767]: I0317 16:01:59.081780 4767 generic.go:334] "Generic (PLEG): container finished" podID="e33e642c-58c5-4783-955a-aba880206b8c" containerID="aeba7e474edd62db80c2ceb1ada7b1450076b2cd0c459187eac2aeca88bd8ca2" exitCode=0 Mar 17 16:01:59 crc kubenswrapper[4767]: I0317 16:01:59.081835 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-nt244" event={"ID":"e33e642c-58c5-4783-955a-aba880206b8c","Type":"ContainerDied","Data":"aeba7e474edd62db80c2ceb1ada7b1450076b2cd0c459187eac2aeca88bd8ca2"} Mar 17 16:01:59 crc kubenswrapper[4767]: I0317 16:01:59.081869 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-nt244" event={"ID":"e33e642c-58c5-4783-955a-aba880206b8c","Type":"ContainerStarted","Data":"9d77e033908c76917742719602639a8404808342751d16050d29a17fe2930f2b"} Mar 17 16:01:59 crc kubenswrapper[4767]: I0317 16:01:59.299427 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 17 16:01:59 crc kubenswrapper[4767]: I0317 16:01:59.351975 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-1" Mar 17 16:01:59 crc kubenswrapper[4767]: I0317 16:01:59.385673 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 17 16:01:59 crc kubenswrapper[4767]: I0317 16:01:59.830949 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-2" Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.102692 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-nt244" event={"ID":"e33e642c-58c5-4783-955a-aba880206b8c","Type":"ContainerStarted","Data":"a40a20831ceca07701fcbe665bef655e088787f63ff45f3755e51936a33be825"} Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.102792 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.108888 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4hdjg" event={"ID":"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa","Type":"ContainerDied","Data":"e2ddeff074fda785b4914c153886c6fe3788507e73308757b73327834debf484"} Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.108935 4767 generic.go:334] "Generic (PLEG): container finished" podID="9d3ec9f9-2461-4c46-afe3-f71f8b956cfa" containerID="e2ddeff074fda785b4914c153886c6fe3788507e73308757b73327834debf484" exitCode=0 Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.145012 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-nt244" podStartSLOduration=3.144982633 podStartE2EDuration="3.144982633s" podCreationTimestamp="2026-03-17 16:01:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:02:00.132064328 +0000 UTC m=+1511.545380385" watchObservedRunningTime="2026-03-17 16:02:00.144982633 +0000 UTC m=+1511.558298680" Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.161124 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562722-prmjs"] Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.163301 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562722-prmjs" Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.165654 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.166510 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.166797 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.178912 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562722-prmjs"] Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.303445 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgm2t\" (UniqueName: \"kubernetes.io/projected/cd1f4000-0e5a-438c-80ad-2805f6f9d772-kube-api-access-vgm2t\") pod \"auto-csr-approver-29562722-prmjs\" (UID: \"cd1f4000-0e5a-438c-80ad-2805f6f9d772\") " pod="openshift-infra/auto-csr-approver-29562722-prmjs" Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.406218 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgm2t\" (UniqueName: \"kubernetes.io/projected/cd1f4000-0e5a-438c-80ad-2805f6f9d772-kube-api-access-vgm2t\") pod \"auto-csr-approver-29562722-prmjs\" (UID: \"cd1f4000-0e5a-438c-80ad-2805f6f9d772\") " pod="openshift-infra/auto-csr-approver-29562722-prmjs" Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.426874 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgm2t\" (UniqueName: \"kubernetes.io/projected/cd1f4000-0e5a-438c-80ad-2805f6f9d772-kube-api-access-vgm2t\") pod \"auto-csr-approver-29562722-prmjs\" (UID: \"cd1f4000-0e5a-438c-80ad-2805f6f9d772\") " pod="openshift-infra/auto-csr-approver-29562722-prmjs" Mar 17 16:02:00 crc kubenswrapper[4767]: I0317 16:02:00.516424 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562722-prmjs" Mar 17 16:02:01 crc kubenswrapper[4767]: I0317 16:02:01.432208 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562722-prmjs"] Mar 17 16:02:01 crc kubenswrapper[4767]: I0317 16:02:01.891022 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4hdjg" Mar 17 16:02:01 crc kubenswrapper[4767]: I0317 16:02:01.985382 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-config-data\") pod \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " Mar 17 16:02:01 crc kubenswrapper[4767]: I0317 16:02:01.985568 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-combined-ca-bundle\") pod \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " Mar 17 16:02:01 crc kubenswrapper[4767]: I0317 16:02:01.985698 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-db-sync-config-data\") pod \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " Mar 17 16:02:01 crc kubenswrapper[4767]: I0317 16:02:01.985740 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r75dp\" (UniqueName: \"kubernetes.io/projected/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-kube-api-access-r75dp\") pod \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\" (UID: \"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa\") " Mar 17 16:02:01 crc kubenswrapper[4767]: I0317 16:02:01.993306 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-kube-api-access-r75dp" (OuterVolumeSpecName: "kube-api-access-r75dp") pod "9d3ec9f9-2461-4c46-afe3-f71f8b956cfa" (UID: "9d3ec9f9-2461-4c46-afe3-f71f8b956cfa"). InnerVolumeSpecName "kube-api-access-r75dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:01 crc kubenswrapper[4767]: I0317 16:02:01.993644 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9d3ec9f9-2461-4c46-afe3-f71f8b956cfa" (UID: "9d3ec9f9-2461-4c46-afe3-f71f8b956cfa"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.021053 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d3ec9f9-2461-4c46-afe3-f71f8b956cfa" (UID: "9d3ec9f9-2461-4c46-afe3-f71f8b956cfa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.051084 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-config-data" (OuterVolumeSpecName: "config-data") pod "9d3ec9f9-2461-4c46-afe3-f71f8b956cfa" (UID: "9d3ec9f9-2461-4c46-afe3-f71f8b956cfa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.088258 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.088306 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.088322 4767 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.088336 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r75dp\" (UniqueName: \"kubernetes.io/projected/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa-kube-api-access-r75dp\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.130580 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4hdjg" event={"ID":"9d3ec9f9-2461-4c46-afe3-f71f8b956cfa","Type":"ContainerDied","Data":"19f74ed0ab1b263431de35de8ce8745096f4bb14acb0ecc174687ae0c8bb7dd5"} Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.130626 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19f74ed0ab1b263431de35de8ce8745096f4bb14acb0ecc174687ae0c8bb7dd5" Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.130690 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4hdjg" Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.140973 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562722-prmjs" event={"ID":"cd1f4000-0e5a-438c-80ad-2805f6f9d772","Type":"ContainerStarted","Data":"e9de950f807bc4ea91d1ff9df96f59904eceaeab91264fca2f6980733c347eeb"} Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.695730 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-nt244"] Mar 17 16:02:02 crc kubenswrapper[4767]: I0317 16:02:02.696113 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-nt244" podUID="e33e642c-58c5-4783-955a-aba880206b8c" containerName="dnsmasq-dns" containerID="cri-o://a40a20831ceca07701fcbe665bef655e088787f63ff45f3755e51936a33be825" gracePeriod=10 Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.044491 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-c7c7p"] Mar 17 16:02:03 crc kubenswrapper[4767]: E0317 16:02:03.077550 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3ec9f9-2461-4c46-afe3-f71f8b956cfa" containerName="glance-db-sync" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.077595 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3ec9f9-2461-4c46-afe3-f71f8b956cfa" containerName="glance-db-sync" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.077891 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3ec9f9-2461-4c46-afe3-f71f8b956cfa" containerName="glance-db-sync" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.079278 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-c7c7p"] Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.079411 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.179279 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-nt244" event={"ID":"e33e642c-58c5-4783-955a-aba880206b8c","Type":"ContainerDied","Data":"a40a20831ceca07701fcbe665bef655e088787f63ff45f3755e51936a33be825"} Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.179288 4767 generic.go:334] "Generic (PLEG): container finished" podID="e33e642c-58c5-4783-955a-aba880206b8c" containerID="a40a20831ceca07701fcbe665bef655e088787f63ff45f3755e51936a33be825" exitCode=0 Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.226524 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qst6s\" (UniqueName: \"kubernetes.io/projected/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-kube-api-access-qst6s\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.226677 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.226701 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.226752 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.226809 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-config\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.227101 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.335322 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.335953 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.336109 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.338240 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.338840 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.339102 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.339503 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-config\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.340028 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-config\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.340144 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.340302 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qst6s\" (UniqueName: \"kubernetes.io/projected/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-kube-api-access-qst6s\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.341197 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.413944 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qst6s\" (UniqueName: \"kubernetes.io/projected/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-kube-api-access-qst6s\") pod \"dnsmasq-dns-74f6bcbc87-c7c7p\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.490038 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-2trqj"] Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.494232 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2trqj" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.520530 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-2trqj"] Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.550530 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.980329 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.991649 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-82a2-account-create-update-q7rxh"] Mar 17 16:02:03 crc kubenswrapper[4767]: E0317 16:02:03.994624 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e33e642c-58c5-4783-955a-aba880206b8c" containerName="dnsmasq-dns" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.994659 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e33e642c-58c5-4783-955a-aba880206b8c" containerName="dnsmasq-dns" Mar 17 16:02:03 crc kubenswrapper[4767]: E0317 16:02:03.994702 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e33e642c-58c5-4783-955a-aba880206b8c" containerName="init" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.994711 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e33e642c-58c5-4783-955a-aba880206b8c" containerName="init" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.994998 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="e33e642c-58c5-4783-955a-aba880206b8c" containerName="dnsmasq-dns" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.996100 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-82a2-account-create-update-q7rxh" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.999549 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c678939f-b738-4c33-a849-175525de0122-operator-scripts\") pod \"cinder-db-create-2trqj\" (UID: \"c678939f-b738-4c33-a849-175525de0122\") " pod="openstack/cinder-db-create-2trqj" Mar 17 16:02:03 crc kubenswrapper[4767]: I0317 16:02:03.999693 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqxrm\" (UniqueName: \"kubernetes.io/projected/c678939f-b738-4c33-a849-175525de0122-kube-api-access-tqxrm\") pod \"cinder-db-create-2trqj\" (UID: \"c678939f-b738-4c33-a849-175525de0122\") " pod="openstack/cinder-db-create-2trqj" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.011412 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.047044 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-82a2-account-create-update-q7rxh"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.081348 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-mdmg7"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.086998 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-mdmg7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.105937 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-nb\") pod \"e33e642c-58c5-4783-955a-aba880206b8c\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.106090 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-config\") pod \"e33e642c-58c5-4783-955a-aba880206b8c\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.106302 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-sb\") pod \"e33e642c-58c5-4783-955a-aba880206b8c\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.106453 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-svc\") pod \"e33e642c-58c5-4783-955a-aba880206b8c\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.106514 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98kkq\" (UniqueName: \"kubernetes.io/projected/e33e642c-58c5-4783-955a-aba880206b8c-kube-api-access-98kkq\") pod \"e33e642c-58c5-4783-955a-aba880206b8c\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.106566 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-swift-storage-0\") pod \"e33e642c-58c5-4783-955a-aba880206b8c\" (UID: \"e33e642c-58c5-4783-955a-aba880206b8c\") " Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.107396 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94877582-ae44-41c9-8a66-1b7d323c71bb-operator-scripts\") pod \"heat-82a2-account-create-update-q7rxh\" (UID: \"94877582-ae44-41c9-8a66-1b7d323c71bb\") " pod="openstack/heat-82a2-account-create-update-q7rxh" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.107529 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c678939f-b738-4c33-a849-175525de0122-operator-scripts\") pod \"cinder-db-create-2trqj\" (UID: \"c678939f-b738-4c33-a849-175525de0122\") " pod="openstack/cinder-db-create-2trqj" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.107591 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqxrm\" (UniqueName: \"kubernetes.io/projected/c678939f-b738-4c33-a849-175525de0122-kube-api-access-tqxrm\") pod \"cinder-db-create-2trqj\" (UID: \"c678939f-b738-4c33-a849-175525de0122\") " pod="openstack/cinder-db-create-2trqj" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.107690 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlb5p\" (UniqueName: \"kubernetes.io/projected/94877582-ae44-41c9-8a66-1b7d323c71bb-kube-api-access-mlb5p\") pod \"heat-82a2-account-create-update-q7rxh\" (UID: \"94877582-ae44-41c9-8a66-1b7d323c71bb\") " pod="openstack/heat-82a2-account-create-update-q7rxh" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.113403 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c678939f-b738-4c33-a849-175525de0122-operator-scripts\") pod \"cinder-db-create-2trqj\" (UID: \"c678939f-b738-4c33-a849-175525de0122\") " pod="openstack/cinder-db-create-2trqj" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.139127 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-0b49-account-create-update-gdff8"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.148711 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b49-account-create-update-gdff8" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.154851 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.164891 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-mdmg7"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.169344 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqxrm\" (UniqueName: \"kubernetes.io/projected/c678939f-b738-4c33-a849-175525de0122-kube-api-access-tqxrm\") pod \"cinder-db-create-2trqj\" (UID: \"c678939f-b738-4c33-a849-175525de0122\") " pod="openstack/cinder-db-create-2trqj" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.171525 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e33e642c-58c5-4783-955a-aba880206b8c-kube-api-access-98kkq" (OuterVolumeSpecName: "kube-api-access-98kkq") pod "e33e642c-58c5-4783-955a-aba880206b8c" (UID: "e33e642c-58c5-4783-955a-aba880206b8c"). InnerVolumeSpecName "kube-api-access-98kkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.172370 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0b49-account-create-update-gdff8"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.216505 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94877582-ae44-41c9-8a66-1b7d323c71bb-operator-scripts\") pod \"heat-82a2-account-create-update-q7rxh\" (UID: \"94877582-ae44-41c9-8a66-1b7d323c71bb\") " pod="openstack/heat-82a2-account-create-update-q7rxh" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.216870 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw4c6\" (UniqueName: \"kubernetes.io/projected/6adfcc51-d4a8-451b-8685-4e822b1be5f9-kube-api-access-bw4c6\") pod \"heat-db-create-mdmg7\" (UID: \"6adfcc51-d4a8-451b-8685-4e822b1be5f9\") " pod="openstack/heat-db-create-mdmg7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.216979 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpsbc\" (UniqueName: \"kubernetes.io/projected/64df007c-c140-48d9-9964-eff020f5d581-kube-api-access-qpsbc\") pod \"cinder-0b49-account-create-update-gdff8\" (UID: \"64df007c-c140-48d9-9964-eff020f5d581\") " pod="openstack/cinder-0b49-account-create-update-gdff8" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.217349 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlb5p\" (UniqueName: \"kubernetes.io/projected/94877582-ae44-41c9-8a66-1b7d323c71bb-kube-api-access-mlb5p\") pod \"heat-82a2-account-create-update-q7rxh\" (UID: \"94877582-ae44-41c9-8a66-1b7d323c71bb\") " pod="openstack/heat-82a2-account-create-update-q7rxh" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.217469 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64df007c-c140-48d9-9964-eff020f5d581-operator-scripts\") pod \"cinder-0b49-account-create-update-gdff8\" (UID: \"64df007c-c140-48d9-9964-eff020f5d581\") " pod="openstack/cinder-0b49-account-create-update-gdff8" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.217598 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6adfcc51-d4a8-451b-8685-4e822b1be5f9-operator-scripts\") pod \"heat-db-create-mdmg7\" (UID: \"6adfcc51-d4a8-451b-8685-4e822b1be5f9\") " pod="openstack/heat-db-create-mdmg7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.217842 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98kkq\" (UniqueName: \"kubernetes.io/projected/e33e642c-58c5-4783-955a-aba880206b8c-kube-api-access-98kkq\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.218811 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94877582-ae44-41c9-8a66-1b7d323c71bb-operator-scripts\") pod \"heat-82a2-account-create-update-q7rxh\" (UID: \"94877582-ae44-41c9-8a66-1b7d323c71bb\") " pod="openstack/heat-82a2-account-create-update-q7rxh" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.241345 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-zbsv7"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.250096 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.255157 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.255382 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9kzs7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.255440 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.255540 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.270297 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlb5p\" (UniqueName: \"kubernetes.io/projected/94877582-ae44-41c9-8a66-1b7d323c71bb-kube-api-access-mlb5p\") pod \"heat-82a2-account-create-update-q7rxh\" (UID: \"94877582-ae44-41c9-8a66-1b7d323c71bb\") " pod="openstack/heat-82a2-account-create-update-q7rxh" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.296023 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-nt244" event={"ID":"e33e642c-58c5-4783-955a-aba880206b8c","Type":"ContainerDied","Data":"9d77e033908c76917742719602639a8404808342751d16050d29a17fe2930f2b"} Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.296238 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-58d46"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.296741 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-nt244" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.298329 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-58d46" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.299947 4767 scope.go:117] "RemoveContainer" containerID="a40a20831ceca07701fcbe665bef655e088787f63ff45f3755e51936a33be825" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.317960 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e33e642c-58c5-4783-955a-aba880206b8c" (UID: "e33e642c-58c5-4783-955a-aba880206b8c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.324124 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64df007c-c140-48d9-9964-eff020f5d581-operator-scripts\") pod \"cinder-0b49-account-create-update-gdff8\" (UID: \"64df007c-c140-48d9-9964-eff020f5d581\") " pod="openstack/cinder-0b49-account-create-update-gdff8" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.324337 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6adfcc51-d4a8-451b-8685-4e822b1be5f9-operator-scripts\") pod \"heat-db-create-mdmg7\" (UID: \"6adfcc51-d4a8-451b-8685-4e822b1be5f9\") " pod="openstack/heat-db-create-mdmg7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.324480 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-config-data\") pod \"keystone-db-sync-zbsv7\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.324550 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw4c6\" (UniqueName: \"kubernetes.io/projected/6adfcc51-d4a8-451b-8685-4e822b1be5f9-kube-api-access-bw4c6\") pod \"heat-db-create-mdmg7\" (UID: \"6adfcc51-d4a8-451b-8685-4e822b1be5f9\") " pod="openstack/heat-db-create-mdmg7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.325999 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64df007c-c140-48d9-9964-eff020f5d581-operator-scripts\") pod \"cinder-0b49-account-create-update-gdff8\" (UID: \"64df007c-c140-48d9-9964-eff020f5d581\") " pod="openstack/cinder-0b49-account-create-update-gdff8" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.326805 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpsbc\" (UniqueName: \"kubernetes.io/projected/64df007c-c140-48d9-9964-eff020f5d581-kube-api-access-qpsbc\") pod \"cinder-0b49-account-create-update-gdff8\" (UID: \"64df007c-c140-48d9-9964-eff020f5d581\") " pod="openstack/cinder-0b49-account-create-update-gdff8" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.326882 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4dq5\" (UniqueName: \"kubernetes.io/projected/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-kube-api-access-w4dq5\") pod \"keystone-db-sync-zbsv7\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.327232 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-combined-ca-bundle\") pod \"keystone-db-sync-zbsv7\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.328400 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.329079 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6adfcc51-d4a8-451b-8685-4e822b1be5f9-operator-scripts\") pod \"heat-db-create-mdmg7\" (UID: \"6adfcc51-d4a8-451b-8685-4e822b1be5f9\") " pod="openstack/heat-db-create-mdmg7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.356898 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-82a2-account-create-update-q7rxh" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.363751 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zbsv7"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.365846 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562722-prmjs" event={"ID":"cd1f4000-0e5a-438c-80ad-2805f6f9d772","Type":"ContainerStarted","Data":"e23ded531abf1800f2b455e0669e8ae5b55542a41b015cf3583366ddff4093ec"} Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.369024 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpsbc\" (UniqueName: \"kubernetes.io/projected/64df007c-c140-48d9-9964-eff020f5d581-kube-api-access-qpsbc\") pod \"cinder-0b49-account-create-update-gdff8\" (UID: \"64df007c-c140-48d9-9964-eff020f5d581\") " pod="openstack/cinder-0b49-account-create-update-gdff8" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.387031 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw4c6\" (UniqueName: \"kubernetes.io/projected/6adfcc51-d4a8-451b-8685-4e822b1be5f9-kube-api-access-bw4c6\") pod \"heat-db-create-mdmg7\" (UID: \"6adfcc51-d4a8-451b-8685-4e822b1be5f9\") " pod="openstack/heat-db-create-mdmg7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.404448 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e33e642c-58c5-4783-955a-aba880206b8c" (UID: "e33e642c-58c5-4783-955a-aba880206b8c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.404484 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-config" (OuterVolumeSpecName: "config") pod "e33e642c-58c5-4783-955a-aba880206b8c" (UID: "e33e642c-58c5-4783-955a-aba880206b8c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.430716 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-mdmg7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.445205 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-config-data\") pod \"keystone-db-sync-zbsv7\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.496935 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76bz7\" (UniqueName: \"kubernetes.io/projected/a2107299-7bc5-4279-a443-7a266e21cec5-kube-api-access-76bz7\") pod \"neutron-db-create-58d46\" (UID: \"a2107299-7bc5-4279-a443-7a266e21cec5\") " pod="openstack/neutron-db-create-58d46" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.497015 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4dq5\" (UniqueName: \"kubernetes.io/projected/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-kube-api-access-w4dq5\") pod \"keystone-db-sync-zbsv7\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.497047 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2107299-7bc5-4279-a443-7a266e21cec5-operator-scripts\") pod \"neutron-db-create-58d46\" (UID: \"a2107299-7bc5-4279-a443-7a266e21cec5\") " pod="openstack/neutron-db-create-58d46" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.497292 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-combined-ca-bundle\") pod \"keystone-db-sync-zbsv7\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.497395 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.497415 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.469602 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-58d46"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.449061 4767 scope.go:117] "RemoveContainer" containerID="aeba7e474edd62db80c2ceb1ada7b1450076b2cd0c459187eac2aeca88bd8ca2" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.450418 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b49-account-create-update-gdff8" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.499463 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-config-data\") pod \"keystone-db-sync-zbsv7\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.469043 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e33e642c-58c5-4783-955a-aba880206b8c" (UID: "e33e642c-58c5-4783-955a-aba880206b8c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.459226 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2trqj" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.522267 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-combined-ca-bundle\") pod \"keystone-db-sync-zbsv7\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.540967 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4dq5\" (UniqueName: \"kubernetes.io/projected/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-kube-api-access-w4dq5\") pod \"keystone-db-sync-zbsv7\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.567983 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e33e642c-58c5-4783-955a-aba880206b8c" (UID: "e33e642c-58c5-4783-955a-aba880206b8c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.609766 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76bz7\" (UniqueName: \"kubernetes.io/projected/a2107299-7bc5-4279-a443-7a266e21cec5-kube-api-access-76bz7\") pod \"neutron-db-create-58d46\" (UID: \"a2107299-7bc5-4279-a443-7a266e21cec5\") " pod="openstack/neutron-db-create-58d46" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.609902 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2107299-7bc5-4279-a443-7a266e21cec5-operator-scripts\") pod \"neutron-db-create-58d46\" (UID: \"a2107299-7bc5-4279-a443-7a266e21cec5\") " pod="openstack/neutron-db-create-58d46" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.611296 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2107299-7bc5-4279-a443-7a266e21cec5-operator-scripts\") pod \"neutron-db-create-58d46\" (UID: \"a2107299-7bc5-4279-a443-7a266e21cec5\") " pod="openstack/neutron-db-create-58d46" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.611717 4767 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.611900 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e33e642c-58c5-4783-955a-aba880206b8c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.629111 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c546-account-create-update-thtgc"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.632720 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c546-account-create-update-thtgc" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.643933 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.656434 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c546-account-create-update-thtgc"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.658189 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76bz7\" (UniqueName: \"kubernetes.io/projected/a2107299-7bc5-4279-a443-7a266e21cec5-kube-api-access-76bz7\") pod \"neutron-db-create-58d46\" (UID: \"a2107299-7bc5-4279-a443-7a266e21cec5\") " pod="openstack/neutron-db-create-58d46" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.666662 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562722-prmjs" podStartSLOduration=3.430376088 podStartE2EDuration="4.666636674s" podCreationTimestamp="2026-03-17 16:02:00 +0000 UTC" firstStartedPulling="2026-03-17 16:02:01.423059536 +0000 UTC m=+1512.836375583" lastFinishedPulling="2026-03-17 16:02:02.659320122 +0000 UTC m=+1514.072636169" observedRunningTime="2026-03-17 16:02:04.403825197 +0000 UTC m=+1515.817141254" watchObservedRunningTime="2026-03-17 16:02:04.666636674 +0000 UTC m=+1516.079952731" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.677782 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rsxsx"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.691641 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rsxsx" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.710498 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rsxsx"] Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.726641 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gx2s\" (UniqueName: \"kubernetes.io/projected/0527b29f-1349-403b-86b4-8003c04e3925-kube-api-access-7gx2s\") pod \"neutron-c546-account-create-update-thtgc\" (UID: \"0527b29f-1349-403b-86b4-8003c04e3925\") " pod="openstack/neutron-c546-account-create-update-thtgc" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.726721 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0527b29f-1349-403b-86b4-8003c04e3925-operator-scripts\") pod \"neutron-c546-account-create-update-thtgc\" (UID: \"0527b29f-1349-403b-86b4-8003c04e3925\") " pod="openstack/neutron-c546-account-create-update-thtgc" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.966129 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gx2s\" (UniqueName: \"kubernetes.io/projected/0527b29f-1349-403b-86b4-8003c04e3925-kube-api-access-7gx2s\") pod \"neutron-c546-account-create-update-thtgc\" (UID: \"0527b29f-1349-403b-86b4-8003c04e3925\") " pod="openstack/neutron-c546-account-create-update-thtgc" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.966209 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0527b29f-1349-403b-86b4-8003c04e3925-operator-scripts\") pod \"neutron-c546-account-create-update-thtgc\" (UID: \"0527b29f-1349-403b-86b4-8003c04e3925\") " pod="openstack/neutron-c546-account-create-update-thtgc" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.966986 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0527b29f-1349-403b-86b4-8003c04e3925-operator-scripts\") pod \"neutron-c546-account-create-update-thtgc\" (UID: \"0527b29f-1349-403b-86b4-8003c04e3925\") " pod="openstack/neutron-c546-account-create-update-thtgc" Mar 17 16:02:04 crc kubenswrapper[4767]: I0317 16:02:04.983440 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-58d46" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:04.987441 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:04.999402 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-883e-account-create-update-qh7zx"] Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.001226 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-883e-account-create-update-qh7zx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.014809 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.014849 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-883e-account-create-update-qh7zx"] Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.029982 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gx2s\" (UniqueName: \"kubernetes.io/projected/0527b29f-1349-403b-86b4-8003c04e3925-kube-api-access-7gx2s\") pod \"neutron-c546-account-create-update-thtgc\" (UID: \"0527b29f-1349-403b-86b4-8003c04e3925\") " pod="openstack/neutron-c546-account-create-update-thtgc" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.060110 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-c7c7p"] Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.083276 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-nt244"] Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.095563 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5m4h\" (UniqueName: \"kubernetes.io/projected/b6d532ab-b620-428e-a0ae-a4793930a703-kube-api-access-h5m4h\") pod \"barbican-db-create-rsxsx\" (UID: \"b6d532ab-b620-428e-a0ae-a4793930a703\") " pod="openstack/barbican-db-create-rsxsx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.096635 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d532ab-b620-428e-a0ae-a4793930a703-operator-scripts\") pod \"barbican-db-create-rsxsx\" (UID: \"b6d532ab-b620-428e-a0ae-a4793930a703\") " pod="openstack/barbican-db-create-rsxsx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.107010 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-nt244"] Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.199022 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr29z\" (UniqueName: \"kubernetes.io/projected/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-kube-api-access-sr29z\") pod \"barbican-883e-account-create-update-qh7zx\" (UID: \"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2\") " pod="openstack/barbican-883e-account-create-update-qh7zx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.203880 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d532ab-b620-428e-a0ae-a4793930a703-operator-scripts\") pod \"barbican-db-create-rsxsx\" (UID: \"b6d532ab-b620-428e-a0ae-a4793930a703\") " pod="openstack/barbican-db-create-rsxsx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.204145 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5m4h\" (UniqueName: \"kubernetes.io/projected/b6d532ab-b620-428e-a0ae-a4793930a703-kube-api-access-h5m4h\") pod \"barbican-db-create-rsxsx\" (UID: \"b6d532ab-b620-428e-a0ae-a4793930a703\") " pod="openstack/barbican-db-create-rsxsx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.204243 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-operator-scripts\") pod \"barbican-883e-account-create-update-qh7zx\" (UID: \"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2\") " pod="openstack/barbican-883e-account-create-update-qh7zx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.220106 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d532ab-b620-428e-a0ae-a4793930a703-operator-scripts\") pod \"barbican-db-create-rsxsx\" (UID: \"b6d532ab-b620-428e-a0ae-a4793930a703\") " pod="openstack/barbican-db-create-rsxsx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.275650 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5m4h\" (UniqueName: \"kubernetes.io/projected/b6d532ab-b620-428e-a0ae-a4793930a703-kube-api-access-h5m4h\") pod \"barbican-db-create-rsxsx\" (UID: \"b6d532ab-b620-428e-a0ae-a4793930a703\") " pod="openstack/barbican-db-create-rsxsx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.308009 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-operator-scripts\") pod \"barbican-883e-account-create-update-qh7zx\" (UID: \"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2\") " pod="openstack/barbican-883e-account-create-update-qh7zx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.308270 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr29z\" (UniqueName: \"kubernetes.io/projected/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-kube-api-access-sr29z\") pod \"barbican-883e-account-create-update-qh7zx\" (UID: \"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2\") " pod="openstack/barbican-883e-account-create-update-qh7zx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.310388 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-operator-scripts\") pod \"barbican-883e-account-create-update-qh7zx\" (UID: \"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2\") " pod="openstack/barbican-883e-account-create-update-qh7zx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.337523 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr29z\" (UniqueName: \"kubernetes.io/projected/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-kube-api-access-sr29z\") pod \"barbican-883e-account-create-update-qh7zx\" (UID: \"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2\") " pod="openstack/barbican-883e-account-create-update-qh7zx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.432004 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e33e642c-58c5-4783-955a-aba880206b8c" path="/var/lib/kubelet/pods/e33e642c-58c5-4783-955a-aba880206b8c/volumes" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.432862 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" event={"ID":"7d4fdc1a-5119-445d-82fd-7f25fd5c879b","Type":"ContainerStarted","Data":"be7af1f32b1e05cebe1c09d9405387a59e0f5af9b49c19244bdda3043ea1f85d"} Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.543284 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-82a2-account-create-update-q7rxh"] Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.550329 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c546-account-create-update-thtgc" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.704277 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rsxsx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.715492 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-883e-account-create-update-qh7zx" Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.877579 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-2trqj"] Mar 17 16:02:05 crc kubenswrapper[4767]: I0317 16:02:05.952048 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-mdmg7"] Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.274890 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0b49-account-create-update-gdff8"] Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.430907 4767 generic.go:334] "Generic (PLEG): container finished" podID="cd1f4000-0e5a-438c-80ad-2805f6f9d772" containerID="e23ded531abf1800f2b455e0669e8ae5b55542a41b015cf3583366ddff4093ec" exitCode=0 Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.431053 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562722-prmjs" event={"ID":"cd1f4000-0e5a-438c-80ad-2805f6f9d772","Type":"ContainerDied","Data":"e23ded531abf1800f2b455e0669e8ae5b55542a41b015cf3583366ddff4093ec"} Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.440394 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-mdmg7" event={"ID":"6adfcc51-d4a8-451b-8685-4e822b1be5f9","Type":"ContainerStarted","Data":"1bd8e4de1a839668ccf959e171b9ea20040c711f606b42ab587083b27bd9fc48"} Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.466127 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-82a2-account-create-update-q7rxh" event={"ID":"94877582-ae44-41c9-8a66-1b7d323c71bb","Type":"ContainerStarted","Data":"618cf3f695ffdadeb94614b37181f82116d21bdff20d659899e46c42ce0adeb9"} Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.496234 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2trqj" event={"ID":"c678939f-b738-4c33-a849-175525de0122","Type":"ContainerStarted","Data":"8e162f4e34ae435092ba34c096e01353860c91f0cd2363b2bc238c2bb53f6a86"} Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.511035 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b49-account-create-update-gdff8" event={"ID":"64df007c-c140-48d9-9964-eff020f5d581","Type":"ContainerStarted","Data":"8f7686af3d1f6d61fefdb74158c8dc4be0aff989394932b738671441a0b1eba1"} Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.527289 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c546-account-create-update-thtgc"] Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.543497 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-58d46"] Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.562386 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zbsv7"] Mar 17 16:02:06 crc kubenswrapper[4767]: W0317 16:02:06.576186 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6b00064_98f4_4f9d_82cb_a6aee03dcbeb.slice/crio-2f59af2910aecffd26a30a9612ea59b69b36d0c9adc3a325962e5ac818220b00 WatchSource:0}: Error finding container 2f59af2910aecffd26a30a9612ea59b69b36d0c9adc3a325962e5ac818220b00: Status 404 returned error can't find the container with id 2f59af2910aecffd26a30a9612ea59b69b36d0c9adc3a325962e5ac818220b00 Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.804780 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rsxsx"] Mar 17 16:02:06 crc kubenswrapper[4767]: I0317 16:02:06.974781 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-883e-account-create-update-qh7zx"] Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.587665 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c546-account-create-update-thtgc" event={"ID":"0527b29f-1349-403b-86b4-8003c04e3925","Type":"ContainerStarted","Data":"5a889d3abc4ff7270493d857f5dc795cb50a42f5bbe73708eae640967edb8669"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.588037 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c546-account-create-update-thtgc" event={"ID":"0527b29f-1349-403b-86b4-8003c04e3925","Type":"ContainerStarted","Data":"3bce5f73b34a896619ec2c5d61d814f485842eb02346824eaba0379f7c5346e9"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.594409 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2trqj" event={"ID":"c678939f-b738-4c33-a849-175525de0122","Type":"ContainerStarted","Data":"a4fc72aebda01e02bd581b2042c114d5ae3b7b185f54c2c57855e9704e2932b8"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.611229 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-883e-account-create-update-qh7zx" event={"ID":"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2","Type":"ContainerStarted","Data":"d0e8afb8e4a04c05eeea84bd1c08a897ea98eb0dc1a8aa418836e6d3e9b015a2"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.611304 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-883e-account-create-update-qh7zx" event={"ID":"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2","Type":"ContainerStarted","Data":"1df9af634467285c8b141628253b3c873ef416e39c5795c0c6bb5de90fd335e3"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.616103 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-58d46" event={"ID":"a2107299-7bc5-4279-a443-7a266e21cec5","Type":"ContainerStarted","Data":"c76e829586e52e8d8891839b90d6ee389c265ed0a0da0268cff0c9cf09ad579e"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.616153 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-58d46" event={"ID":"a2107299-7bc5-4279-a443-7a266e21cec5","Type":"ContainerStarted","Data":"a66ade11d310bfaae93ec9fe424f7a9cbe31c11a6c275492a922c6eb95f2713e"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.621302 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-mdmg7" event={"ID":"6adfcc51-d4a8-451b-8685-4e822b1be5f9","Type":"ContainerStarted","Data":"07ab6f0deb52c5b01b03c33b2995e9757ba273c5e3b14e9cdb1d31e8cf2065bc"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.629301 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b49-account-create-update-gdff8" event={"ID":"64df007c-c140-48d9-9964-eff020f5d581","Type":"ContainerStarted","Data":"72a4f5442998b9192efa01ca5f0c3807d15f253aeefc3e81e59f202eae4f5c41"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.637365 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rsxsx" event={"ID":"b6d532ab-b620-428e-a0ae-a4793930a703","Type":"ContainerStarted","Data":"a00da337f62baf20a19d907fed47f45f34e8040f172fcb2e0b78fabd9ceddb08"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.637418 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rsxsx" event={"ID":"b6d532ab-b620-428e-a0ae-a4793930a703","Type":"ContainerStarted","Data":"358c67061a1808214878d8d0a10826e163364a8643228c98e8d4514631c1c4a1"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.638716 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-2trqj" podStartSLOduration=4.638691666 podStartE2EDuration="4.638691666s" podCreationTimestamp="2026-03-17 16:02:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:02:07.623351247 +0000 UTC m=+1519.036667304" watchObservedRunningTime="2026-03-17 16:02:07.638691666 +0000 UTC m=+1519.052007713" Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.654266 4767 generic.go:334] "Generic (PLEG): container finished" podID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" containerID="a60de704a8164f69ec89d8cb9100672d84bc03253f0d8eed00f04ef221a1ec35" exitCode=0 Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.654392 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" event={"ID":"7d4fdc1a-5119-445d-82fd-7f25fd5c879b","Type":"ContainerDied","Data":"a60de704a8164f69ec89d8cb9100672d84bc03253f0d8eed00f04ef221a1ec35"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.663006 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zbsv7" event={"ID":"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb","Type":"ContainerStarted","Data":"2f59af2910aecffd26a30a9612ea59b69b36d0c9adc3a325962e5ac818220b00"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.665773 4767 generic.go:334] "Generic (PLEG): container finished" podID="94877582-ae44-41c9-8a66-1b7d323c71bb" containerID="bf9587e52cd4d776af85f2888f638c71ef1d56c858ac4293357a4b8e18541c40" exitCode=0 Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.666188 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-82a2-account-create-update-q7rxh" event={"ID":"94877582-ae44-41c9-8a66-1b7d323c71bb","Type":"ContainerDied","Data":"bf9587e52cd4d776af85f2888f638c71ef1d56c858ac4293357a4b8e18541c40"} Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.671107 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-883e-account-create-update-qh7zx" podStartSLOduration=3.671090081 podStartE2EDuration="3.671090081s" podCreationTimestamp="2026-03-17 16:02:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:02:07.64405261 +0000 UTC m=+1519.057368657" watchObservedRunningTime="2026-03-17 16:02:07.671090081 +0000 UTC m=+1519.084406128" Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.698604 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-58d46" podStartSLOduration=3.698576475 podStartE2EDuration="3.698576475s" podCreationTimestamp="2026-03-17 16:02:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:02:07.676570718 +0000 UTC m=+1519.089886775" watchObservedRunningTime="2026-03-17 16:02:07.698576475 +0000 UTC m=+1519.111892532" Mar 17 16:02:07 crc kubenswrapper[4767]: I0317 16:02:07.804768 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-rsxsx" podStartSLOduration=3.80473935 podStartE2EDuration="3.80473935s" podCreationTimestamp="2026-03-17 16:02:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:02:07.780164564 +0000 UTC m=+1519.193480621" watchObservedRunningTime="2026-03-17 16:02:07.80473935 +0000 UTC m=+1519.218055397" Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.175499 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562722-prmjs" Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.198695 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgm2t\" (UniqueName: \"kubernetes.io/projected/cd1f4000-0e5a-438c-80ad-2805f6f9d772-kube-api-access-vgm2t\") pod \"cd1f4000-0e5a-438c-80ad-2805f6f9d772\" (UID: \"cd1f4000-0e5a-438c-80ad-2805f6f9d772\") " Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.210891 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd1f4000-0e5a-438c-80ad-2805f6f9d772-kube-api-access-vgm2t" (OuterVolumeSpecName: "kube-api-access-vgm2t") pod "cd1f4000-0e5a-438c-80ad-2805f6f9d772" (UID: "cd1f4000-0e5a-438c-80ad-2805f6f9d772"). InnerVolumeSpecName "kube-api-access-vgm2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.302225 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgm2t\" (UniqueName: \"kubernetes.io/projected/cd1f4000-0e5a-438c-80ad-2805f6f9d772-kube-api-access-vgm2t\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.650370 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562716-54ssn"] Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.675689 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562716-54ssn"] Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.690090 4767 generic.go:334] "Generic (PLEG): container finished" podID="6adfcc51-d4a8-451b-8685-4e822b1be5f9" containerID="07ab6f0deb52c5b01b03c33b2995e9757ba273c5e3b14e9cdb1d31e8cf2065bc" exitCode=0 Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.690400 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-mdmg7" event={"ID":"6adfcc51-d4a8-451b-8685-4e822b1be5f9","Type":"ContainerDied","Data":"07ab6f0deb52c5b01b03c33b2995e9757ba273c5e3b14e9cdb1d31e8cf2065bc"} Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.693007 4767 generic.go:334] "Generic (PLEG): container finished" podID="c678939f-b738-4c33-a849-175525de0122" containerID="a4fc72aebda01e02bd581b2042c114d5ae3b7b185f54c2c57855e9704e2932b8" exitCode=0 Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.693206 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2trqj" event={"ID":"c678939f-b738-4c33-a849-175525de0122","Type":"ContainerDied","Data":"a4fc72aebda01e02bd581b2042c114d5ae3b7b185f54c2c57855e9704e2932b8"} Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.696086 4767 generic.go:334] "Generic (PLEG): container finished" podID="64df007c-c140-48d9-9964-eff020f5d581" containerID="72a4f5442998b9192efa01ca5f0c3807d15f253aeefc3e81e59f202eae4f5c41" exitCode=0 Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.696331 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b49-account-create-update-gdff8" event={"ID":"64df007c-c140-48d9-9964-eff020f5d581","Type":"ContainerDied","Data":"72a4f5442998b9192efa01ca5f0c3807d15f253aeefc3e81e59f202eae4f5c41"} Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.699582 4767 generic.go:334] "Generic (PLEG): container finished" podID="b6d532ab-b620-428e-a0ae-a4793930a703" containerID="a00da337f62baf20a19d907fed47f45f34e8040f172fcb2e0b78fabd9ceddb08" exitCode=0 Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.699916 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rsxsx" event={"ID":"b6d532ab-b620-428e-a0ae-a4793930a703","Type":"ContainerDied","Data":"a00da337f62baf20a19d907fed47f45f34e8040f172fcb2e0b78fabd9ceddb08"} Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.703638 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" event={"ID":"7d4fdc1a-5119-445d-82fd-7f25fd5c879b","Type":"ContainerStarted","Data":"65995c95f1119cef59d260897bd801a022b8dec2d4889addcfd1373665067005"} Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.705063 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.715747 4767 generic.go:334] "Generic (PLEG): container finished" podID="0527b29f-1349-403b-86b4-8003c04e3925" containerID="5a889d3abc4ff7270493d857f5dc795cb50a42f5bbe73708eae640967edb8669" exitCode=0 Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.716211 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c546-account-create-update-thtgc" event={"ID":"0527b29f-1349-403b-86b4-8003c04e3925","Type":"ContainerDied","Data":"5a889d3abc4ff7270493d857f5dc795cb50a42f5bbe73708eae640967edb8669"} Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.725394 4767 generic.go:334] "Generic (PLEG): container finished" podID="ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2" containerID="d0e8afb8e4a04c05eeea84bd1c08a897ea98eb0dc1a8aa418836e6d3e9b015a2" exitCode=0 Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.725483 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-883e-account-create-update-qh7zx" event={"ID":"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2","Type":"ContainerDied","Data":"d0e8afb8e4a04c05eeea84bd1c08a897ea98eb0dc1a8aa418836e6d3e9b015a2"} Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.728818 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562722-prmjs" event={"ID":"cd1f4000-0e5a-438c-80ad-2805f6f9d772","Type":"ContainerDied","Data":"e9de950f807bc4ea91d1ff9df96f59904eceaeab91264fca2f6980733c347eeb"} Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.728943 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9de950f807bc4ea91d1ff9df96f59904eceaeab91264fca2f6980733c347eeb" Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.729062 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562722-prmjs" Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.737143 4767 generic.go:334] "Generic (PLEG): container finished" podID="a2107299-7bc5-4279-a443-7a266e21cec5" containerID="c76e829586e52e8d8891839b90d6ee389c265ed0a0da0268cff0c9cf09ad579e" exitCode=0 Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.737486 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-58d46" event={"ID":"a2107299-7bc5-4279-a443-7a266e21cec5","Type":"ContainerDied","Data":"c76e829586e52e8d8891839b90d6ee389c265ed0a0da0268cff0c9cf09ad579e"} Mar 17 16:02:08 crc kubenswrapper[4767]: I0317 16:02:08.744345 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" podStartSLOduration=6.744318379 podStartE2EDuration="6.744318379s" podCreationTimestamp="2026-03-17 16:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:02:08.735444632 +0000 UTC m=+1520.148760699" watchObservedRunningTime="2026-03-17 16:02:08.744318379 +0000 UTC m=+1520.157634426" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.261162 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-mdmg7" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.384740 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d7c1cce-3729-4157-9f4e-275c46c6e436" path="/var/lib/kubelet/pods/9d7c1cce-3729-4157-9f4e-275c46c6e436/volumes" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.421621 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw4c6\" (UniqueName: \"kubernetes.io/projected/6adfcc51-d4a8-451b-8685-4e822b1be5f9-kube-api-access-bw4c6\") pod \"6adfcc51-d4a8-451b-8685-4e822b1be5f9\" (UID: \"6adfcc51-d4a8-451b-8685-4e822b1be5f9\") " Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.422087 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6adfcc51-d4a8-451b-8685-4e822b1be5f9-operator-scripts\") pod \"6adfcc51-d4a8-451b-8685-4e822b1be5f9\" (UID: \"6adfcc51-d4a8-451b-8685-4e822b1be5f9\") " Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.423064 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6adfcc51-d4a8-451b-8685-4e822b1be5f9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6adfcc51-d4a8-451b-8685-4e822b1be5f9" (UID: "6adfcc51-d4a8-451b-8685-4e822b1be5f9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.434375 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6adfcc51-d4a8-451b-8685-4e822b1be5f9-kube-api-access-bw4c6" (OuterVolumeSpecName: "kube-api-access-bw4c6") pod "6adfcc51-d4a8-451b-8685-4e822b1be5f9" (UID: "6adfcc51-d4a8-451b-8685-4e822b1be5f9"). InnerVolumeSpecName "kube-api-access-bw4c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.524495 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw4c6\" (UniqueName: \"kubernetes.io/projected/6adfcc51-d4a8-451b-8685-4e822b1be5f9-kube-api-access-bw4c6\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.524528 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6adfcc51-d4a8-451b-8685-4e822b1be5f9-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.595263 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.595515 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.603057 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.609246 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-82a2-account-create-update-q7rxh" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.615734 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b49-account-create-update-gdff8" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.641329 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c546-account-create-update-thtgc" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.737858 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gx2s\" (UniqueName: \"kubernetes.io/projected/0527b29f-1349-403b-86b4-8003c04e3925-kube-api-access-7gx2s\") pod \"0527b29f-1349-403b-86b4-8003c04e3925\" (UID: \"0527b29f-1349-403b-86b4-8003c04e3925\") " Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.737928 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlb5p\" (UniqueName: \"kubernetes.io/projected/94877582-ae44-41c9-8a66-1b7d323c71bb-kube-api-access-mlb5p\") pod \"94877582-ae44-41c9-8a66-1b7d323c71bb\" (UID: \"94877582-ae44-41c9-8a66-1b7d323c71bb\") " Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.738132 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0527b29f-1349-403b-86b4-8003c04e3925-operator-scripts\") pod \"0527b29f-1349-403b-86b4-8003c04e3925\" (UID: \"0527b29f-1349-403b-86b4-8003c04e3925\") " Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.739563 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0527b29f-1349-403b-86b4-8003c04e3925-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0527b29f-1349-403b-86b4-8003c04e3925" (UID: "0527b29f-1349-403b-86b4-8003c04e3925"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.741355 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64df007c-c140-48d9-9964-eff020f5d581-operator-scripts\") pod \"64df007c-c140-48d9-9964-eff020f5d581\" (UID: \"64df007c-c140-48d9-9964-eff020f5d581\") " Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.741449 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94877582-ae44-41c9-8a66-1b7d323c71bb-operator-scripts\") pod \"94877582-ae44-41c9-8a66-1b7d323c71bb\" (UID: \"94877582-ae44-41c9-8a66-1b7d323c71bb\") " Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.741489 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpsbc\" (UniqueName: \"kubernetes.io/projected/64df007c-c140-48d9-9964-eff020f5d581-kube-api-access-qpsbc\") pod \"64df007c-c140-48d9-9964-eff020f5d581\" (UID: \"64df007c-c140-48d9-9964-eff020f5d581\") " Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.741890 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94877582-ae44-41c9-8a66-1b7d323c71bb-kube-api-access-mlb5p" (OuterVolumeSpecName: "kube-api-access-mlb5p") pod "94877582-ae44-41c9-8a66-1b7d323c71bb" (UID: "94877582-ae44-41c9-8a66-1b7d323c71bb"). InnerVolumeSpecName "kube-api-access-mlb5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.743138 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94877582-ae44-41c9-8a66-1b7d323c71bb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "94877582-ae44-41c9-8a66-1b7d323c71bb" (UID: "94877582-ae44-41c9-8a66-1b7d323c71bb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.743502 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64df007c-c140-48d9-9964-eff020f5d581-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "64df007c-c140-48d9-9964-eff020f5d581" (UID: "64df007c-c140-48d9-9964-eff020f5d581"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.746294 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlb5p\" (UniqueName: \"kubernetes.io/projected/94877582-ae44-41c9-8a66-1b7d323c71bb-kube-api-access-mlb5p\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.746725 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0527b29f-1349-403b-86b4-8003c04e3925-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.746774 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64df007c-c140-48d9-9964-eff020f5d581-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.746788 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94877582-ae44-41c9-8a66-1b7d323c71bb-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.750058 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0527b29f-1349-403b-86b4-8003c04e3925-kube-api-access-7gx2s" (OuterVolumeSpecName: "kube-api-access-7gx2s") pod "0527b29f-1349-403b-86b4-8003c04e3925" (UID: "0527b29f-1349-403b-86b4-8003c04e3925"). InnerVolumeSpecName "kube-api-access-7gx2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.760924 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64df007c-c140-48d9-9964-eff020f5d581-kube-api-access-qpsbc" (OuterVolumeSpecName: "kube-api-access-qpsbc") pod "64df007c-c140-48d9-9964-eff020f5d581" (UID: "64df007c-c140-48d9-9964-eff020f5d581"). InnerVolumeSpecName "kube-api-access-qpsbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.772900 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-mdmg7" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.773084 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-mdmg7" event={"ID":"6adfcc51-d4a8-451b-8685-4e822b1be5f9","Type":"ContainerDied","Data":"1bd8e4de1a839668ccf959e171b9ea20040c711f606b42ab587083b27bd9fc48"} Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.773497 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bd8e4de1a839668ccf959e171b9ea20040c711f606b42ab587083b27bd9fc48" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.782233 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-82a2-account-create-update-q7rxh" event={"ID":"94877582-ae44-41c9-8a66-1b7d323c71bb","Type":"ContainerDied","Data":"618cf3f695ffdadeb94614b37181f82116d21bdff20d659899e46c42ce0adeb9"} Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.782280 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="618cf3f695ffdadeb94614b37181f82116d21bdff20d659899e46c42ce0adeb9" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.782349 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-82a2-account-create-update-q7rxh" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.797942 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c546-account-create-update-thtgc" event={"ID":"0527b29f-1349-403b-86b4-8003c04e3925","Type":"ContainerDied","Data":"3bce5f73b34a896619ec2c5d61d814f485842eb02346824eaba0379f7c5346e9"} Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.797978 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c546-account-create-update-thtgc" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.797999 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bce5f73b34a896619ec2c5d61d814f485842eb02346824eaba0379f7c5346e9" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.803444 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0b49-account-create-update-gdff8" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.803989 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0b49-account-create-update-gdff8" event={"ID":"64df007c-c140-48d9-9964-eff020f5d581","Type":"ContainerDied","Data":"8f7686af3d1f6d61fefdb74158c8dc4be0aff989394932b738671441a0b1eba1"} Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.804067 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f7686af3d1f6d61fefdb74158c8dc4be0aff989394932b738671441a0b1eba1" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.854045 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gx2s\" (UniqueName: \"kubernetes.io/projected/0527b29f-1349-403b-86b4-8003c04e3925-kube-api-access-7gx2s\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:09 crc kubenswrapper[4767]: I0317 16:02:09.854083 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpsbc\" (UniqueName: \"kubernetes.io/projected/64df007c-c140-48d9-9964-eff020f5d581-kube-api-access-qpsbc\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:13 crc kubenswrapper[4767]: I0317 16:02:13.558373 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:13 crc kubenswrapper[4767]: I0317 16:02:13.637623 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tt884"] Mar 17 16:02:13 crc kubenswrapper[4767]: I0317 16:02:13.637911 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-tt884" podUID="788da75a-4303-4878-a388-217e6beee11b" containerName="dnsmasq-dns" containerID="cri-o://f926d0748efcabdaa65f2e4500ba9d42c0ab10d5476dd9dd4e281cd161bfe90b" gracePeriod=10 Mar 17 16:02:13 crc kubenswrapper[4767]: I0317 16:02:13.884949 4767 generic.go:334] "Generic (PLEG): container finished" podID="788da75a-4303-4878-a388-217e6beee11b" containerID="f926d0748efcabdaa65f2e4500ba9d42c0ab10d5476dd9dd4e281cd161bfe90b" exitCode=0 Mar 17 16:02:13 crc kubenswrapper[4767]: I0317 16:02:13.885057 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tt884" event={"ID":"788da75a-4303-4878-a388-217e6beee11b","Type":"ContainerDied","Data":"f926d0748efcabdaa65f2e4500ba9d42c0ab10d5476dd9dd4e281cd161bfe90b"} Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.726934 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-58d46" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.743000 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rsxsx" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.785014 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-883e-account-create-update-qh7zx" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.816280 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2trqj" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.847428 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2107299-7bc5-4279-a443-7a266e21cec5-operator-scripts\") pod \"a2107299-7bc5-4279-a443-7a266e21cec5\" (UID: \"a2107299-7bc5-4279-a443-7a266e21cec5\") " Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.848211 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2107299-7bc5-4279-a443-7a266e21cec5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a2107299-7bc5-4279-a443-7a266e21cec5" (UID: "a2107299-7bc5-4279-a443-7a266e21cec5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.848437 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76bz7\" (UniqueName: \"kubernetes.io/projected/a2107299-7bc5-4279-a443-7a266e21cec5-kube-api-access-76bz7\") pod \"a2107299-7bc5-4279-a443-7a266e21cec5\" (UID: \"a2107299-7bc5-4279-a443-7a266e21cec5\") " Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.849795 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2107299-7bc5-4279-a443-7a266e21cec5-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.855544 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2107299-7bc5-4279-a443-7a266e21cec5-kube-api-access-76bz7" (OuterVolumeSpecName: "kube-api-access-76bz7") pod "a2107299-7bc5-4279-a443-7a266e21cec5" (UID: "a2107299-7bc5-4279-a443-7a266e21cec5"). InnerVolumeSpecName "kube-api-access-76bz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.911694 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2trqj" event={"ID":"c678939f-b738-4c33-a849-175525de0122","Type":"ContainerDied","Data":"8e162f4e34ae435092ba34c096e01353860c91f0cd2363b2bc238c2bb53f6a86"} Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.911735 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e162f4e34ae435092ba34c096e01353860c91f0cd2363b2bc238c2bb53f6a86" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.912326 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2trqj" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.914152 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-883e-account-create-update-qh7zx" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.914149 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-883e-account-create-update-qh7zx" event={"ID":"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2","Type":"ContainerDied","Data":"1df9af634467285c8b141628253b3c873ef416e39c5795c0c6bb5de90fd335e3"} Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.914737 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1df9af634467285c8b141628253b3c873ef416e39c5795c0c6bb5de90fd335e3" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.914979 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.915834 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rsxsx" event={"ID":"b6d532ab-b620-428e-a0ae-a4793930a703","Type":"ContainerDied","Data":"358c67061a1808214878d8d0a10826e163364a8643228c98e8d4514631c1c4a1"} Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.915863 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="358c67061a1808214878d8d0a10826e163364a8643228c98e8d4514631c1c4a1" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.916088 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rsxsx" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.919231 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tt884" event={"ID":"788da75a-4303-4878-a388-217e6beee11b","Type":"ContainerDied","Data":"aa3eb1ef2027a57c8a0a5526e34a64782c7b406e6bb553cfd3327ad757a8f18a"} Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.919298 4767 scope.go:117] "RemoveContainer" containerID="f926d0748efcabdaa65f2e4500ba9d42c0ab10d5476dd9dd4e281cd161bfe90b" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.919423 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tt884" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.921980 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zbsv7" event={"ID":"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb","Type":"ContainerStarted","Data":"a0f1af2de4965844a935ea59d92271222ff3b15093bb1f72eedb28d97f9e7e2b"} Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.924390 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-58d46" event={"ID":"a2107299-7bc5-4279-a443-7a266e21cec5","Type":"ContainerDied","Data":"a66ade11d310bfaae93ec9fe424f7a9cbe31c11a6c275492a922c6eb95f2713e"} Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.924430 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a66ade11d310bfaae93ec9fe424f7a9cbe31c11a6c275492a922c6eb95f2713e" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.924485 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-58d46" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.951279 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5m4h\" (UniqueName: \"kubernetes.io/projected/b6d532ab-b620-428e-a0ae-a4793930a703-kube-api-access-h5m4h\") pod \"b6d532ab-b620-428e-a0ae-a4793930a703\" (UID: \"b6d532ab-b620-428e-a0ae-a4793930a703\") " Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.951494 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-operator-scripts\") pod \"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2\" (UID: \"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2\") " Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.951626 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqxrm\" (UniqueName: \"kubernetes.io/projected/c678939f-b738-4c33-a849-175525de0122-kube-api-access-tqxrm\") pod \"c678939f-b738-4c33-a849-175525de0122\" (UID: \"c678939f-b738-4c33-a849-175525de0122\") " Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.951652 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c678939f-b738-4c33-a849-175525de0122-operator-scripts\") pod \"c678939f-b738-4c33-a849-175525de0122\" (UID: \"c678939f-b738-4c33-a849-175525de0122\") " Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.951676 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d532ab-b620-428e-a0ae-a4793930a703-operator-scripts\") pod \"b6d532ab-b620-428e-a0ae-a4793930a703\" (UID: \"b6d532ab-b620-428e-a0ae-a4793930a703\") " Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.951788 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr29z\" (UniqueName: \"kubernetes.io/projected/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-kube-api-access-sr29z\") pod \"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2\" (UID: \"ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2\") " Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.952350 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76bz7\" (UniqueName: \"kubernetes.io/projected/a2107299-7bc5-4279-a443-7a266e21cec5-kube-api-access-76bz7\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.953262 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6d532ab-b620-428e-a0ae-a4793930a703-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b6d532ab-b620-428e-a0ae-a4793930a703" (UID: "b6d532ab-b620-428e-a0ae-a4793930a703"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.953341 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c678939f-b738-4c33-a849-175525de0122-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c678939f-b738-4c33-a849-175525de0122" (UID: "c678939f-b738-4c33-a849-175525de0122"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.953819 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2" (UID: "ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.970279 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d532ab-b620-428e-a0ae-a4793930a703-kube-api-access-h5m4h" (OuterVolumeSpecName: "kube-api-access-h5m4h") pod "b6d532ab-b620-428e-a0ae-a4793930a703" (UID: "b6d532ab-b620-428e-a0ae-a4793930a703"). InnerVolumeSpecName "kube-api-access-h5m4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.970368 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c678939f-b738-4c33-a849-175525de0122-kube-api-access-tqxrm" (OuterVolumeSpecName: "kube-api-access-tqxrm") pod "c678939f-b738-4c33-a849-175525de0122" (UID: "c678939f-b738-4c33-a849-175525de0122"). InnerVolumeSpecName "kube-api-access-tqxrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.975055 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-zbsv7" podStartSLOduration=3.109369381 podStartE2EDuration="10.975030932s" podCreationTimestamp="2026-03-17 16:02:04 +0000 UTC" firstStartedPulling="2026-03-17 16:02:06.606933787 +0000 UTC m=+1518.020249834" lastFinishedPulling="2026-03-17 16:02:14.472595338 +0000 UTC m=+1525.885911385" observedRunningTime="2026-03-17 16:02:14.97119378 +0000 UTC m=+1526.384509847" watchObservedRunningTime="2026-03-17 16:02:14.975030932 +0000 UTC m=+1526.388346969" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.978274 4767 scope.go:117] "RemoveContainer" containerID="9ede4241b1f554851e9b2d432b9fc848e9d31ef9c6c6dff53102be9491b7f53b" Mar 17 16:02:14 crc kubenswrapper[4767]: I0317 16:02:14.979516 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-kube-api-access-sr29z" (OuterVolumeSpecName: "kube-api-access-sr29z") pod "ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2" (UID: "ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2"). InnerVolumeSpecName "kube-api-access-sr29z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.054581 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-nb\") pod \"788da75a-4303-4878-a388-217e6beee11b\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.054707 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbpr6\" (UniqueName: \"kubernetes.io/projected/788da75a-4303-4878-a388-217e6beee11b-kube-api-access-jbpr6\") pod \"788da75a-4303-4878-a388-217e6beee11b\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.054851 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-config\") pod \"788da75a-4303-4878-a388-217e6beee11b\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.054956 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-sb\") pod \"788da75a-4303-4878-a388-217e6beee11b\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.055032 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-dns-svc\") pod \"788da75a-4303-4878-a388-217e6beee11b\" (UID: \"788da75a-4303-4878-a388-217e6beee11b\") " Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.055760 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.055781 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqxrm\" (UniqueName: \"kubernetes.io/projected/c678939f-b738-4c33-a849-175525de0122-kube-api-access-tqxrm\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.055797 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c678939f-b738-4c33-a849-175525de0122-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.055810 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6d532ab-b620-428e-a0ae-a4793930a703-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.055822 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr29z\" (UniqueName: \"kubernetes.io/projected/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2-kube-api-access-sr29z\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.055834 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5m4h\" (UniqueName: \"kubernetes.io/projected/b6d532ab-b620-428e-a0ae-a4793930a703-kube-api-access-h5m4h\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.062999 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788da75a-4303-4878-a388-217e6beee11b-kube-api-access-jbpr6" (OuterVolumeSpecName: "kube-api-access-jbpr6") pod "788da75a-4303-4878-a388-217e6beee11b" (UID: "788da75a-4303-4878-a388-217e6beee11b"). InnerVolumeSpecName "kube-api-access-jbpr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.112410 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "788da75a-4303-4878-a388-217e6beee11b" (UID: "788da75a-4303-4878-a388-217e6beee11b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.113190 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "788da75a-4303-4878-a388-217e6beee11b" (UID: "788da75a-4303-4878-a388-217e6beee11b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.119771 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "788da75a-4303-4878-a388-217e6beee11b" (UID: "788da75a-4303-4878-a388-217e6beee11b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.130677 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-config" (OuterVolumeSpecName: "config") pod "788da75a-4303-4878-a388-217e6beee11b" (UID: "788da75a-4303-4878-a388-217e6beee11b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.157828 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.157863 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbpr6\" (UniqueName: \"kubernetes.io/projected/788da75a-4303-4878-a388-217e6beee11b-kube-api-access-jbpr6\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.157877 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.157886 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.157896 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/788da75a-4303-4878-a388-217e6beee11b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.288503 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tt884"] Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.297701 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tt884"] Mar 17 16:02:15 crc kubenswrapper[4767]: I0317 16:02:15.369134 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="788da75a-4303-4878-a388-217e6beee11b" path="/var/lib/kubelet/pods/788da75a-4303-4878-a388-217e6beee11b/volumes" Mar 17 16:02:17 crc kubenswrapper[4767]: I0317 16:02:17.996681 4767 generic.go:334] "Generic (PLEG): container finished" podID="f6b00064-98f4-4f9d-82cb-a6aee03dcbeb" containerID="a0f1af2de4965844a935ea59d92271222ff3b15093bb1f72eedb28d97f9e7e2b" exitCode=0 Mar 17 16:02:17 crc kubenswrapper[4767]: I0317 16:02:17.996785 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zbsv7" event={"ID":"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb","Type":"ContainerDied","Data":"a0f1af2de4965844a935ea59d92271222ff3b15093bb1f72eedb28d97f9e7e2b"} Mar 17 16:02:19 crc kubenswrapper[4767]: I0317 16:02:19.468493 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:19 crc kubenswrapper[4767]: I0317 16:02:19.650585 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4dq5\" (UniqueName: \"kubernetes.io/projected/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-kube-api-access-w4dq5\") pod \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " Mar 17 16:02:19 crc kubenswrapper[4767]: I0317 16:02:19.651079 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-config-data\") pod \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " Mar 17 16:02:19 crc kubenswrapper[4767]: I0317 16:02:19.651221 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-combined-ca-bundle\") pod \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\" (UID: \"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb\") " Mar 17 16:02:19 crc kubenswrapper[4767]: I0317 16:02:19.663750 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-kube-api-access-w4dq5" (OuterVolumeSpecName: "kube-api-access-w4dq5") pod "f6b00064-98f4-4f9d-82cb-a6aee03dcbeb" (UID: "f6b00064-98f4-4f9d-82cb-a6aee03dcbeb"). InnerVolumeSpecName "kube-api-access-w4dq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:19 crc kubenswrapper[4767]: I0317 16:02:19.685675 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6b00064-98f4-4f9d-82cb-a6aee03dcbeb" (UID: "f6b00064-98f4-4f9d-82cb-a6aee03dcbeb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:19 crc kubenswrapper[4767]: I0317 16:02:19.715842 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-config-data" (OuterVolumeSpecName: "config-data") pod "f6b00064-98f4-4f9d-82cb-a6aee03dcbeb" (UID: "f6b00064-98f4-4f9d-82cb-a6aee03dcbeb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:19 crc kubenswrapper[4767]: I0317 16:02:19.754815 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:19 crc kubenswrapper[4767]: I0317 16:02:19.754859 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4dq5\" (UniqueName: \"kubernetes.io/projected/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-kube-api-access-w4dq5\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:19 crc kubenswrapper[4767]: I0317 16:02:19.754870 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.107411 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zbsv7" event={"ID":"f6b00064-98f4-4f9d-82cb-a6aee03dcbeb","Type":"ContainerDied","Data":"2f59af2910aecffd26a30a9612ea59b69b36d0c9adc3a325962e5ac818220b00"} Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.107701 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f59af2910aecffd26a30a9612ea59b69b36d0c9adc3a325962e5ac818220b00" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.107737 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zbsv7" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.346050 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8w6w6"] Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347072 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2107299-7bc5-4279-a443-7a266e21cec5" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347094 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2107299-7bc5-4279-a443-7a266e21cec5" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347106 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6adfcc51-d4a8-451b-8685-4e822b1be5f9" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347113 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6adfcc51-d4a8-451b-8685-4e822b1be5f9" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347131 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788da75a-4303-4878-a388-217e6beee11b" containerName="dnsmasq-dns" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347139 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="788da75a-4303-4878-a388-217e6beee11b" containerName="dnsmasq-dns" Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347148 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c678939f-b738-4c33-a849-175525de0122" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347153 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c678939f-b738-4c33-a849-175525de0122" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347188 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0527b29f-1349-403b-86b4-8003c04e3925" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347194 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0527b29f-1349-403b-86b4-8003c04e3925" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347206 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d532ab-b620-428e-a0ae-a4793930a703" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347212 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d532ab-b620-428e-a0ae-a4793930a703" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347226 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94877582-ae44-41c9-8a66-1b7d323c71bb" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347232 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="94877582-ae44-41c9-8a66-1b7d323c71bb" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347242 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64df007c-c140-48d9-9964-eff020f5d581" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347251 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="64df007c-c140-48d9-9964-eff020f5d581" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347272 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347278 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347288 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd1f4000-0e5a-438c-80ad-2805f6f9d772" containerName="oc" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347293 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd1f4000-0e5a-438c-80ad-2805f6f9d772" containerName="oc" Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347314 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788da75a-4303-4878-a388-217e6beee11b" containerName="init" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347321 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="788da75a-4303-4878-a388-217e6beee11b" containerName="init" Mar 17 16:02:20 crc kubenswrapper[4767]: E0317 16:02:20.347331 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b00064-98f4-4f9d-82cb-a6aee03dcbeb" containerName="keystone-db-sync" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347338 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b00064-98f4-4f9d-82cb-a6aee03dcbeb" containerName="keystone-db-sync" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347565 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0527b29f-1349-403b-86b4-8003c04e3925" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347582 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="788da75a-4303-4878-a388-217e6beee11b" containerName="dnsmasq-dns" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347591 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347597 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b00064-98f4-4f9d-82cb-a6aee03dcbeb" containerName="keystone-db-sync" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347607 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c678939f-b738-4c33-a849-175525de0122" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347622 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd1f4000-0e5a-438c-80ad-2805f6f9d772" containerName="oc" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347630 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2107299-7bc5-4279-a443-7a266e21cec5" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347643 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="94877582-ae44-41c9-8a66-1b7d323c71bb" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347654 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="6adfcc51-d4a8-451b-8685-4e822b1be5f9" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347667 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="64df007c-c140-48d9-9964-eff020f5d581" containerName="mariadb-account-create-update" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.347675 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d532ab-b620-428e-a0ae-a4793930a703" containerName="mariadb-database-create" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.349980 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.387739 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8w6w6"] Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.395190 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcpxx\" (UniqueName: \"kubernetes.io/projected/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-kube-api-access-tcpxx\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.395327 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-svc\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.395463 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-config\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.395523 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.395807 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.395936 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.416466 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fxjjm"] Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.418690 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.422593 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9kzs7" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.422813 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.422940 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.423097 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.441689 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.507902 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcpxx\" (UniqueName: \"kubernetes.io/projected/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-kube-api-access-tcpxx\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.509218 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-svc\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.509342 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-combined-ca-bundle\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.509413 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-credential-keys\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.509608 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-config\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.509640 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.509713 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-scripts\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.509880 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-config-data\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.509983 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.510022 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7dpc\" (UniqueName: \"kubernetes.io/projected/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-kube-api-access-f7dpc\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.510080 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-fernet-keys\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.510111 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.510853 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-svc\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.511562 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-config\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.512127 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.513119 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.530296 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.600016 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcpxx\" (UniqueName: \"kubernetes.io/projected/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-kube-api-access-tcpxx\") pod \"dnsmasq-dns-847c4cc679-8w6w6\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.605666 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fxjjm"] Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.614081 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-fernet-keys\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.614272 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-combined-ca-bundle\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.614304 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-credential-keys\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.614377 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-scripts\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.614434 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-config-data\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.614484 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7dpc\" (UniqueName: \"kubernetes.io/projected/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-kube-api-access-f7dpc\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.619975 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-combined-ca-bundle\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.623049 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-credential-keys\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.630318 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-scripts\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.644966 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-fernet-keys\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.650376 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-config-data\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.665098 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7dpc\" (UniqueName: \"kubernetes.io/projected/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-kube-api-access-f7dpc\") pod \"keystone-bootstrap-fxjjm\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.677256 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.682425 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-657sp"] Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.685040 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-657sp" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.700163 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-kmsxr" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.719321 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-combined-ca-bundle\") pod \"heat-db-sync-657sp\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " pod="openstack/heat-db-sync-657sp" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.719388 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-config-data\") pod \"heat-db-sync-657sp\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " pod="openstack/heat-db-sync-657sp" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.719529 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8kxb\" (UniqueName: \"kubernetes.io/projected/5e6a688d-7438-45a1-9aa7-b7042485163f-kube-api-access-z8kxb\") pod \"heat-db-sync-657sp\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " pod="openstack/heat-db-sync-657sp" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.729064 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.739249 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-657sp"] Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.767120 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.827618 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-combined-ca-bundle\") pod \"heat-db-sync-657sp\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " pod="openstack/heat-db-sync-657sp" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.828015 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-config-data\") pod \"heat-db-sync-657sp\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " pod="openstack/heat-db-sync-657sp" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.828188 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8kxb\" (UniqueName: \"kubernetes.io/projected/5e6a688d-7438-45a1-9aa7-b7042485163f-kube-api-access-z8kxb\") pod \"heat-db-sync-657sp\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " pod="openstack/heat-db-sync-657sp" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.846038 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-config-data\") pod \"heat-db-sync-657sp\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " pod="openstack/heat-db-sync-657sp" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.857549 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-combined-ca-bundle\") pod \"heat-db-sync-657sp\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " pod="openstack/heat-db-sync-657sp" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.888361 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-pdxns"] Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.890868 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.899795 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.900063 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7rfgj" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.906127 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.939827 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8kxb\" (UniqueName: \"kubernetes.io/projected/5e6a688d-7438-45a1-9aa7-b7042485163f-kube-api-access-z8kxb\") pod \"heat-db-sync-657sp\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " pod="openstack/heat-db-sync-657sp" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.943162 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jmsj\" (UniqueName: \"kubernetes.io/projected/712803ce-6e47-40a6-bf5c-84e8aea748e4-kube-api-access-6jmsj\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.943291 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-config-data\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.943447 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-scripts\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.943549 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-combined-ca-bundle\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.943597 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-db-sync-config-data\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:20 crc kubenswrapper[4767]: I0317 16:02:20.943623 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/712803ce-6e47-40a6-bf5c-84e8aea748e4-etc-machine-id\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.204695 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-combined-ca-bundle\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.204787 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-db-sync-config-data\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.204825 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/712803ce-6e47-40a6-bf5c-84e8aea748e4-etc-machine-id\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.204947 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jmsj\" (UniqueName: \"kubernetes.io/projected/712803ce-6e47-40a6-bf5c-84e8aea748e4-kube-api-access-6jmsj\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.205028 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-config-data\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.205234 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-scripts\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.217391 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-657sp" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.218420 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-scripts\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.218517 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/712803ce-6e47-40a6-bf5c-84e8aea748e4-etc-machine-id\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.220523 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-pdxns"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.229324 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-combined-ca-bundle\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.233995 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-db-sync-config-data\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.237837 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-config-data\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.264461 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8w6w6"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.282906 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-h7hgh"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.285802 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.321644 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jmsj\" (UniqueName: \"kubernetes.io/projected/712803ce-6e47-40a6-bf5c-84e8aea748e4-kube-api-access-6jmsj\") pod \"cinder-db-sync-pdxns\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.351759 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-xj9pj"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.354796 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.362982 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rnsnz" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.367824 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.407574 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-h7hgh"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.407625 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xj9pj"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.414460 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.414748 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmb6t\" (UniqueName: \"kubernetes.io/projected/4fd01229-44e5-40e9-ba3c-d672309ca045-kube-api-access-hmb6t\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.414799 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.414830 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.414996 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-config\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.415343 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.421121 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-gm7c4"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.423384 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.434004 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.434304 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.434432 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-2gvvz" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.446520 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-gm7c4"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.476271 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-mzst5"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.478424 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.490070 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-mzst5"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.501317 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.520830 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.522196 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.522457 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5glpn\" (UniqueName: \"kubernetes.io/projected/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-kube-api-access-5glpn\") pod \"barbican-db-sync-xj9pj\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.522623 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-combined-ca-bundle\") pod \"barbican-db-sync-xj9pj\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.522761 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmb6t\" (UniqueName: \"kubernetes.io/projected/4fd01229-44e5-40e9-ba3c-d672309ca045-kube-api-access-hmb6t\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.522884 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.523015 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.523407 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-config\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.523502 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-combined-ca-bundle\") pod \"neutron-db-sync-gm7c4\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.523555 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-config\") pod \"neutron-db-sync-gm7c4\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.523634 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz6rx\" (UniqueName: \"kubernetes.io/projected/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-kube-api-access-zz6rx\") pod \"neutron-db-sync-gm7c4\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.523697 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-db-sync-config-data\") pod \"barbican-db-sync-xj9pj\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.523726 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.522242 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bjvk2" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.522303 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.524655 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.525345 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.525795 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.531592 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-config\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.573536 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmb6t\" (UniqueName: \"kubernetes.io/projected/4fd01229-44e5-40e9-ba3c-d672309ca045-kube-api-access-hmb6t\") pod \"dnsmasq-dns-785d8bcb8c-h7hgh\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.587092 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pdxns" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.625187 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.629230 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.637057 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-combined-ca-bundle\") pod \"neutron-db-sync-gm7c4\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.637130 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-config\") pod \"neutron-db-sync-gm7c4\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.637194 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz6rx\" (UniqueName: \"kubernetes.io/projected/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-kube-api-access-zz6rx\") pod \"neutron-db-sync-gm7c4\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.637228 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-db-sync-config-data\") pod \"barbican-db-sync-xj9pj\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.637260 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-combined-ca-bundle\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.637400 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-config-data\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.637424 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5glpn\" (UniqueName: \"kubernetes.io/projected/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-kube-api-access-5glpn\") pod \"barbican-db-sync-xj9pj\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.637461 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmtvx\" (UniqueName: \"kubernetes.io/projected/c5b58621-27a1-479d-a90c-aa96400e1116-kube-api-access-bmtvx\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.637505 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-combined-ca-bundle\") pod \"barbican-db-sync-xj9pj\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.637585 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-scripts\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.637636 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5b58621-27a1-479d-a90c-aa96400e1116-logs\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.638683 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-hp85d" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.638915 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.639097 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.639439 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.658381 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.671587 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.712197 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.715278 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.722347 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.722580 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.722859 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-combined-ca-bundle\") pod \"barbican-db-sync-xj9pj\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.725083 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-config\") pod \"neutron-db-sync-gm7c4\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.726108 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-combined-ca-bundle\") pod \"neutron-db-sync-gm7c4\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.741693 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5glpn\" (UniqueName: \"kubernetes.io/projected/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-kube-api-access-5glpn\") pod \"barbican-db-sync-xj9pj\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.742427 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5b58621-27a1-479d-a90c-aa96400e1116-logs\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.742619 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-config-data\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.742798 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.742944 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9x5x\" (UniqueName: \"kubernetes.io/projected/a3c2b25b-eec8-4883-80de-f10237d9697d-kube-api-access-k9x5x\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.743095 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-combined-ca-bundle\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.743271 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-logs\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.743421 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.743563 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-scripts\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.743715 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.743834 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.743946 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-config-data\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.744070 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmtvx\" (UniqueName: \"kubernetes.io/projected/c5b58621-27a1-479d-a90c-aa96400e1116-kube-api-access-bmtvx\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.744283 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-scripts\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.746932 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5b58621-27a1-479d-a90c-aa96400e1116-logs\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.750737 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-db-sync-config-data\") pod \"barbican-db-sync-xj9pj\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.764935 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz6rx\" (UniqueName: \"kubernetes.io/projected/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-kube-api-access-zz6rx\") pod \"neutron-db-sync-gm7c4\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.773507 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-config-data\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.773736 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-combined-ca-bundle\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.775593 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-scripts\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.776977 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmtvx\" (UniqueName: \"kubernetes.io/projected/c5b58621-27a1-479d-a90c-aa96400e1116-kube-api-access-bmtvx\") pod \"placement-db-sync-mzst5\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.781259 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.853874 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-config-data\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.853938 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.855568 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9x5x\" (UniqueName: \"kubernetes.io/projected/a3c2b25b-eec8-4883-80de-f10237d9697d-kube-api-access-k9x5x\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.855773 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-logs\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.855938 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.856032 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-scripts\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.856182 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.856231 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.856460 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-logs\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.857066 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.863981 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.864033 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/683557e62de6f02370abe27ab62c9a425f8a422dda40ecee534d661837e8035d/globalmount\"" pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.879281 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-config-data\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.882486 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.890127 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-scripts\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.894954 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.897694 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.910622 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9x5x\" (UniqueName: \"kubernetes.io/projected/a3c2b25b-eec8-4883-80de-f10237d9697d-kube-api-access-k9x5x\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.958944 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.959378 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.959403 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-scripts\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.959606 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95z5m\" (UniqueName: \"kubernetes.io/projected/32c935bf-9481-4636-802c-3ad3b9082531-kube-api-access-95z5m\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.959634 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.959652 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-logs\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.959668 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-config-data\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:21 crc kubenswrapper[4767]: I0317 16:02:21.959698 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.152636 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.152803 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.152850 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-scripts\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.153304 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95z5m\" (UniqueName: \"kubernetes.io/projected/32c935bf-9481-4636-802c-3ad3b9082531-kube-api-access-95z5m\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.153390 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.153437 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-logs\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.153470 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-config-data\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.155009 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.157352 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.162983 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-logs\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.167308 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.168138 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.186440 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.186873 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5f7c8528f2df23f78ebd936b9a4e66a32c7cd4122d28a5f082f69cd7529229c4/globalmount\"" pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.204783 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.207974 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-scripts\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.211542 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-config-data\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.224044 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95z5m\" (UniqueName: \"kubernetes.io/projected/32c935bf-9481-4636-802c-3ad3b9082531-kube-api-access-95z5m\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.349909 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.388993 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mzst5" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.408258 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.451607 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.468760 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.473396 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.473511 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-scripts\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.473567 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25njk\" (UniqueName: \"kubernetes.io/projected/d17172fd-996e-496e-92b5-9cd19edc65f4-kube-api-access-25njk\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.473645 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-run-httpd\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.473814 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.473859 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-config-data\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.474671 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-log-httpd\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.475138 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.480627 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.512680 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.513518 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.578195 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.578281 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-config-data\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.578350 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-log-httpd\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.578402 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.579820 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-scripts\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.579920 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25njk\" (UniqueName: \"kubernetes.io/projected/d17172fd-996e-496e-92b5-9cd19edc65f4-kube-api-access-25njk\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.580056 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-run-httpd\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.582575 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-run-httpd\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.582819 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fxjjm"] Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.582926 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-log-httpd\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.587225 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-config-data\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.591458 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.595652 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-scripts\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.600763 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.604444 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25njk\" (UniqueName: \"kubernetes.io/projected/d17172fd-996e-496e-92b5-9cd19edc65f4-kube-api-access-25njk\") pod \"ceilometer-0\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.786976 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.828339 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-657sp"] Mar 17 16:02:22 crc kubenswrapper[4767]: W0317 16:02:22.841487 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e6a688d_7438_45a1_9aa7_b7042485163f.slice/crio-8ee5d42cb494ad0292cece8e63bd267fae73fd7589f072159cae76fd6785fe0d WatchSource:0}: Error finding container 8ee5d42cb494ad0292cece8e63bd267fae73fd7589f072159cae76fd6785fe0d: Status 404 returned error can't find the container with id 8ee5d42cb494ad0292cece8e63bd267fae73fd7589f072159cae76fd6785fe0d Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.846822 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:02:22 crc kubenswrapper[4767]: I0317 16:02:22.848289 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8w6w6"] Mar 17 16:02:23 crc kubenswrapper[4767]: I0317 16:02:23.496461 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" event={"ID":"24b48278-cd39-4c5e-b2e2-97ef9f609fa0","Type":"ContainerStarted","Data":"fe4c3531518e5da6b3883929bde60ac9fd814f69903e976178125dd9ebbe4625"} Mar 17 16:02:23 crc kubenswrapper[4767]: I0317 16:02:23.500726 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-657sp" event={"ID":"5e6a688d-7438-45a1-9aa7-b7042485163f","Type":"ContainerStarted","Data":"8ee5d42cb494ad0292cece8e63bd267fae73fd7589f072159cae76fd6785fe0d"} Mar 17 16:02:23 crc kubenswrapper[4767]: I0317 16:02:23.508214 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fxjjm" event={"ID":"4ee82178-eb34-4b0e-9fa1-ffa7b786431d","Type":"ContainerStarted","Data":"0dab081d8810ac517e5792d0ceeee372d5ab8873ed4c7434b2d8caee8d2deda1"} Mar 17 16:02:23 crc kubenswrapper[4767]: I0317 16:02:23.508269 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fxjjm" event={"ID":"4ee82178-eb34-4b0e-9fa1-ffa7b786431d","Type":"ContainerStarted","Data":"3609b35d210f32bce8b66bb1b7cb0f921b1e89f1b0a7a7a9997cc0911c5f8607"} Mar 17 16:02:23 crc kubenswrapper[4767]: I0317 16:02:23.581956 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fxjjm" podStartSLOduration=3.581930383 podStartE2EDuration="3.581930383s" podCreationTimestamp="2026-03-17 16:02:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:02:23.570102458 +0000 UTC m=+1534.983418505" watchObservedRunningTime="2026-03-17 16:02:23.581930383 +0000 UTC m=+1534.995246430" Mar 17 16:02:23 crc kubenswrapper[4767]: I0317 16:02:23.785759 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:02:23 crc kubenswrapper[4767]: I0317 16:02:23.826424 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-mzst5"] Mar 17 16:02:23 crc kubenswrapper[4767]: I0317 16:02:23.881512 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-pdxns"] Mar 17 16:02:23 crc kubenswrapper[4767]: W0317 16:02:23.899650 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod712803ce_6e47_40a6_bf5c_84e8aea748e4.slice/crio-c22c642712671e1561ca6caa6e4eba8e87c35ff51dce05cbc0c14b6a680aa990 WatchSource:0}: Error finding container c22c642712671e1561ca6caa6e4eba8e87c35ff51dce05cbc0c14b6a680aa990: Status 404 returned error can't find the container with id c22c642712671e1561ca6caa6e4eba8e87c35ff51dce05cbc0c14b6a680aa990 Mar 17 16:02:23 crc kubenswrapper[4767]: I0317 16:02:23.913409 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-gm7c4"] Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:23.997668 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xj9pj"] Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.123435 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-h7hgh"] Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.174970 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.566988 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gm7c4" event={"ID":"958adc8c-3c1c-4b35-8cc0-fa7a3b600842","Type":"ContainerStarted","Data":"e4ce5c7ada7a6cfb87e8eeae2a4274ae320237b8835e7247161ddc3add750393"} Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.582153 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xj9pj" event={"ID":"850bce7d-74fa-4ead-b6c2-fcf21c3083ec","Type":"ContainerStarted","Data":"4477daea19710bf9ccf749a74478f8df1c9640d0aba65c0ad1fade1b7305c3b3"} Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.629793 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pdxns" event={"ID":"712803ce-6e47-40a6-bf5c-84e8aea748e4","Type":"ContainerStarted","Data":"c22c642712671e1561ca6caa6e4eba8e87c35ff51dce05cbc0c14b6a680aa990"} Mar 17 16:02:24 crc kubenswrapper[4767]: W0317 16:02:24.635293 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd17172fd_996e_496e_92b5_9cd19edc65f4.slice/crio-b341537ccb4b136de4ee6b2f73d91bb94b0972196b2cc9404f3b92ace4936702 WatchSource:0}: Error finding container b341537ccb4b136de4ee6b2f73d91bb94b0972196b2cc9404f3b92ace4936702: Status 404 returned error can't find the container with id b341537ccb4b136de4ee6b2f73d91bb94b0972196b2cc9404f3b92ace4936702 Mar 17 16:02:24 crc kubenswrapper[4767]: W0317 16:02:24.648099 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3c2b25b_eec8_4883_80de_f10237d9697d.slice/crio-39929050b6cd1c12eab8ac758fb9e75e46b337f0eb9a904d94bf7188aa637b5d WatchSource:0}: Error finding container 39929050b6cd1c12eab8ac758fb9e75e46b337f0eb9a904d94bf7188aa637b5d: Status 404 returned error can't find the container with id 39929050b6cd1c12eab8ac758fb9e75e46b337f0eb9a904d94bf7188aa637b5d Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.651708 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mzst5" event={"ID":"c5b58621-27a1-479d-a90c-aa96400e1116","Type":"ContainerStarted","Data":"9886db56504484e4251c813478a7c5d8201378f98cbc61258596d6ebd57e344b"} Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.660539 4767 generic.go:334] "Generic (PLEG): container finished" podID="24b48278-cd39-4c5e-b2e2-97ef9f609fa0" containerID="fecb8d7098ea5ea1bc4ae07a3589219a434efe94315781247fc4560388c0fd4f" exitCode=0 Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.660700 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" event={"ID":"24b48278-cd39-4c5e-b2e2-97ef9f609fa0","Type":"ContainerDied","Data":"fecb8d7098ea5ea1bc4ae07a3589219a434efe94315781247fc4560388c0fd4f"} Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.800813 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" event={"ID":"4fd01229-44e5-40e9-ba3c-d672309ca045","Type":"ContainerStarted","Data":"5b09d51e9aae6c4d80a24cb7ce2265e67243b148da3da69429bc3f1468141ac4"} Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.811283 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.828491 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:02:24 crc kubenswrapper[4767]: I0317 16:02:24.916592 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:02:25 crc kubenswrapper[4767]: I0317 16:02:25.062655 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:02:25 crc kubenswrapper[4767]: W0317 16:02:25.104826 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32c935bf_9481_4636_802c_3ad3b9082531.slice/crio-c1c832781d89720410ae811bef7951cb32893c83ca15cde8e83a46d77e68b338 WatchSource:0}: Error finding container c1c832781d89720410ae811bef7951cb32893c83ca15cde8e83a46d77e68b338: Status 404 returned error can't find the container with id c1c832781d89720410ae811bef7951cb32893c83ca15cde8e83a46d77e68b338 Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.313461 4767 generic.go:334] "Generic (PLEG): container finished" podID="4fd01229-44e5-40e9-ba3c-d672309ca045" containerID="5cfd31f76b2969652897d667470f38b4cfd75254f1d21c04073a4cb121e9fd73" exitCode=0 Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.316399 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" event={"ID":"4fd01229-44e5-40e9-ba3c-d672309ca045","Type":"ContainerDied","Data":"5cfd31f76b2969652897d667470f38b4cfd75254f1d21c04073a4cb121e9fd73"} Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.338684 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3c2b25b-eec8-4883-80de-f10237d9697d","Type":"ContainerStarted","Data":"39929050b6cd1c12eab8ac758fb9e75e46b337f0eb9a904d94bf7188aa637b5d"} Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.376159 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gm7c4" event={"ID":"958adc8c-3c1c-4b35-8cc0-fa7a3b600842","Type":"ContainerStarted","Data":"207a85d187102e1a1442134ad96bbcd58337d2da80439bd774ba9ac839b6831e"} Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.454492 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"32c935bf-9481-4636-802c-3ad3b9082531","Type":"ContainerStarted","Data":"c1c832781d89720410ae811bef7951cb32893c83ca15cde8e83a46d77e68b338"} Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.502818 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-gm7c4" podStartSLOduration=6.502787935 podStartE2EDuration="6.502787935s" podCreationTimestamp="2026-03-17 16:02:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:02:26.443202955 +0000 UTC m=+1537.856519012" watchObservedRunningTime="2026-03-17 16:02:26.502787935 +0000 UTC m=+1537.916103972" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.503931 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17172fd-996e-496e-92b5-9cd19edc65f4","Type":"ContainerStarted","Data":"b341537ccb4b136de4ee6b2f73d91bb94b0972196b2cc9404f3b92ace4936702"} Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.554705 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.642086 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-sb\") pod \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.642218 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcpxx\" (UniqueName: \"kubernetes.io/projected/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-kube-api-access-tcpxx\") pod \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.642288 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-swift-storage-0\") pod \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.642313 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-config\") pod \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.642334 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-nb\") pod \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.642367 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-svc\") pod \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\" (UID: \"24b48278-cd39-4c5e-b2e2-97ef9f609fa0\") " Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.690615 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "24b48278-cd39-4c5e-b2e2-97ef9f609fa0" (UID: "24b48278-cd39-4c5e-b2e2-97ef9f609fa0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.701266 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-kube-api-access-tcpxx" (OuterVolumeSpecName: "kube-api-access-tcpxx") pod "24b48278-cd39-4c5e-b2e2-97ef9f609fa0" (UID: "24b48278-cd39-4c5e-b2e2-97ef9f609fa0"). InnerVolumeSpecName "kube-api-access-tcpxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.731864 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "24b48278-cd39-4c5e-b2e2-97ef9f609fa0" (UID: "24b48278-cd39-4c5e-b2e2-97ef9f609fa0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.732435 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "24b48278-cd39-4c5e-b2e2-97ef9f609fa0" (UID: "24b48278-cd39-4c5e-b2e2-97ef9f609fa0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.746042 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-config" (OuterVolumeSpecName: "config") pod "24b48278-cd39-4c5e-b2e2-97ef9f609fa0" (UID: "24b48278-cd39-4c5e-b2e2-97ef9f609fa0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.751898 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "24b48278-cd39-4c5e-b2e2-97ef9f609fa0" (UID: "24b48278-cd39-4c5e-b2e2-97ef9f609fa0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.752916 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.753051 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcpxx\" (UniqueName: \"kubernetes.io/projected/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-kube-api-access-tcpxx\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.753151 4767 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.753288 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.753557 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:26 crc kubenswrapper[4767]: I0317 16:02:26.753644 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24b48278-cd39-4c5e-b2e2-97ef9f609fa0-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:27 crc kubenswrapper[4767]: I0317 16:02:27.589735 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" Mar 17 16:02:27 crc kubenswrapper[4767]: I0317 16:02:27.620231 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:27 crc kubenswrapper[4767]: I0317 16:02:27.620273 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-8w6w6" event={"ID":"24b48278-cd39-4c5e-b2e2-97ef9f609fa0","Type":"ContainerDied","Data":"fe4c3531518e5da6b3883929bde60ac9fd814f69903e976178125dd9ebbe4625"} Mar 17 16:02:27 crc kubenswrapper[4767]: I0317 16:02:27.620395 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" event={"ID":"4fd01229-44e5-40e9-ba3c-d672309ca045","Type":"ContainerStarted","Data":"a58959ac80935fea8311b60a5217d0be5b31501cd9709bf2e61a38d6101605aa"} Mar 17 16:02:27 crc kubenswrapper[4767]: I0317 16:02:27.620442 4767 scope.go:117] "RemoveContainer" containerID="fecb8d7098ea5ea1bc4ae07a3589219a434efe94315781247fc4560388c0fd4f" Mar 17 16:02:27 crc kubenswrapper[4767]: I0317 16:02:27.719895 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" podStartSLOduration=7.71986527 podStartE2EDuration="7.71986527s" podCreationTimestamp="2026-03-17 16:02:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:02:27.623289251 +0000 UTC m=+1539.036605318" watchObservedRunningTime="2026-03-17 16:02:27.71986527 +0000 UTC m=+1539.133181317" Mar 17 16:02:27 crc kubenswrapper[4767]: I0317 16:02:27.783989 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8w6w6"] Mar 17 16:02:27 crc kubenswrapper[4767]: I0317 16:02:27.802589 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8w6w6"] Mar 17 16:02:28 crc kubenswrapper[4767]: I0317 16:02:28.627756 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3c2b25b-eec8-4883-80de-f10237d9697d","Type":"ContainerStarted","Data":"5632c8f1a2ead0b8063f52708160719115c6d7fc19cc764d22f41e8d1708f1f3"} Mar 17 16:02:28 crc kubenswrapper[4767]: I0317 16:02:28.636904 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"32c935bf-9481-4636-802c-3ad3b9082531","Type":"ContainerStarted","Data":"089c8fbd067c8329f240985ab1e4b647aafe09a409bbf0b7b9643b3e3dea65c1"} Mar 17 16:02:29 crc kubenswrapper[4767]: I0317 16:02:29.397342 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24b48278-cd39-4c5e-b2e2-97ef9f609fa0" path="/var/lib/kubelet/pods/24b48278-cd39-4c5e-b2e2-97ef9f609fa0/volumes" Mar 17 16:02:29 crc kubenswrapper[4767]: I0317 16:02:29.987103 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a3c2b25b-eec8-4883-80de-f10237d9697d" containerName="glance-log" containerID="cri-o://5632c8f1a2ead0b8063f52708160719115c6d7fc19cc764d22f41e8d1708f1f3" gracePeriod=30 Mar 17 16:02:29 crc kubenswrapper[4767]: I0317 16:02:29.987329 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3c2b25b-eec8-4883-80de-f10237d9697d","Type":"ContainerStarted","Data":"574213173cd2f9deec1dca3468ae8fac70759231f0f10274544925ad9f23671c"} Mar 17 16:02:29 crc kubenswrapper[4767]: I0317 16:02:29.987368 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a3c2b25b-eec8-4883-80de-f10237d9697d" containerName="glance-httpd" containerID="cri-o://574213173cd2f9deec1dca3468ae8fac70759231f0f10274544925ad9f23671c" gracePeriod=30 Mar 17 16:02:30 crc kubenswrapper[4767]: I0317 16:02:30.135191 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.135147074 podStartE2EDuration="10.135147074s" podCreationTimestamp="2026-03-17 16:02:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:02:30.119101355 +0000 UTC m=+1541.532417422" watchObservedRunningTime="2026-03-17 16:02:30.135147074 +0000 UTC m=+1541.548463121" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.007166 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"32c935bf-9481-4636-802c-3ad3b9082531","Type":"ContainerStarted","Data":"40df73eddb1de8a40bda91f5552300a5e2d28eb41e9f7230bbc8d0996d13ae7a"} Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.007888 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="32c935bf-9481-4636-802c-3ad3b9082531" containerName="glance-log" containerID="cri-o://089c8fbd067c8329f240985ab1e4b647aafe09a409bbf0b7b9643b3e3dea65c1" gracePeriod=30 Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.008305 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="32c935bf-9481-4636-802c-3ad3b9082531" containerName="glance-httpd" containerID="cri-o://40df73eddb1de8a40bda91f5552300a5e2d28eb41e9f7230bbc8d0996d13ae7a" gracePeriod=30 Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.025518 4767 generic.go:334] "Generic (PLEG): container finished" podID="a3c2b25b-eec8-4883-80de-f10237d9697d" containerID="574213173cd2f9deec1dca3468ae8fac70759231f0f10274544925ad9f23671c" exitCode=143 Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.025556 4767 generic.go:334] "Generic (PLEG): container finished" podID="a3c2b25b-eec8-4883-80de-f10237d9697d" containerID="5632c8f1a2ead0b8063f52708160719115c6d7fc19cc764d22f41e8d1708f1f3" exitCode=143 Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.025588 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3c2b25b-eec8-4883-80de-f10237d9697d","Type":"ContainerDied","Data":"574213173cd2f9deec1dca3468ae8fac70759231f0f10274544925ad9f23671c"} Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.025633 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3c2b25b-eec8-4883-80de-f10237d9697d","Type":"ContainerDied","Data":"5632c8f1a2ead0b8063f52708160719115c6d7fc19cc764d22f41e8d1708f1f3"} Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.043123 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.043093775 podStartE2EDuration="11.043093775s" podCreationTimestamp="2026-03-17 16:02:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:02:31.036752186 +0000 UTC m=+1542.450068233" watchObservedRunningTime="2026-03-17 16:02:31.043093775 +0000 UTC m=+1542.456409822" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.406792 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.566196 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-public-tls-certs\") pod \"a3c2b25b-eec8-4883-80de-f10237d9697d\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.566260 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-combined-ca-bundle\") pod \"a3c2b25b-eec8-4883-80de-f10237d9697d\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.566443 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-config-data\") pod \"a3c2b25b-eec8-4883-80de-f10237d9697d\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.566486 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-logs\") pod \"a3c2b25b-eec8-4883-80de-f10237d9697d\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.566530 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-scripts\") pod \"a3c2b25b-eec8-4883-80de-f10237d9697d\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.566697 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-httpd-run\") pod \"a3c2b25b-eec8-4883-80de-f10237d9697d\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.566829 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9x5x\" (UniqueName: \"kubernetes.io/projected/a3c2b25b-eec8-4883-80de-f10237d9697d-kube-api-access-k9x5x\") pod \"a3c2b25b-eec8-4883-80de-f10237d9697d\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.567033 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"a3c2b25b-eec8-4883-80de-f10237d9697d\" (UID: \"a3c2b25b-eec8-4883-80de-f10237d9697d\") " Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.571394 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a3c2b25b-eec8-4883-80de-f10237d9697d" (UID: "a3c2b25b-eec8-4883-80de-f10237d9697d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.576697 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-logs" (OuterVolumeSpecName: "logs") pod "a3c2b25b-eec8-4883-80de-f10237d9697d" (UID: "a3c2b25b-eec8-4883-80de-f10237d9697d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.609865 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-scripts" (OuterVolumeSpecName: "scripts") pod "a3c2b25b-eec8-4883-80de-f10237d9697d" (UID: "a3c2b25b-eec8-4883-80de-f10237d9697d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.620910 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3c2b25b-eec8-4883-80de-f10237d9697d-kube-api-access-k9x5x" (OuterVolumeSpecName: "kube-api-access-k9x5x") pod "a3c2b25b-eec8-4883-80de-f10237d9697d" (UID: "a3c2b25b-eec8-4883-80de-f10237d9697d"). InnerVolumeSpecName "kube-api-access-k9x5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.644441 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b" (OuterVolumeSpecName: "glance") pod "a3c2b25b-eec8-4883-80de-f10237d9697d" (UID: "a3c2b25b-eec8-4883-80de-f10237d9697d"). InnerVolumeSpecName "pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.660370 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.914555 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9x5x\" (UniqueName: \"kubernetes.io/projected/a3c2b25b-eec8-4883-80de-f10237d9697d-kube-api-access-k9x5x\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.914937 4767 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") on node \"crc\" " Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.914954 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.914973 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.914984 4767 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3c2b25b-eec8-4883-80de-f10237d9697d-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:31 crc kubenswrapper[4767]: I0317 16:02:31.931975 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a3c2b25b-eec8-4883-80de-f10237d9697d" (UID: "a3c2b25b-eec8-4883-80de-f10237d9697d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.000908 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3c2b25b-eec8-4883-80de-f10237d9697d" (UID: "a3c2b25b-eec8-4883-80de-f10237d9697d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.021525 4767 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.021705 4767 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b") on node "crc" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.027697 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-c7c7p"] Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.027956 4767 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.028031 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.028101 4767 reconciler_common.go:293] "Volume detached for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.028050 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" podUID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" containerName="dnsmasq-dns" containerID="cri-o://65995c95f1119cef59d260897bd801a022b8dec2d4889addcfd1373665067005" gracePeriod=10 Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.039623 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-config-data" (OuterVolumeSpecName: "config-data") pod "a3c2b25b-eec8-4883-80de-f10237d9697d" (UID: "a3c2b25b-eec8-4883-80de-f10237d9697d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.131705 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c2b25b-eec8-4883-80de-f10237d9697d-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.145346 4767 generic.go:334] "Generic (PLEG): container finished" podID="32c935bf-9481-4636-802c-3ad3b9082531" containerID="40df73eddb1de8a40bda91f5552300a5e2d28eb41e9f7230bbc8d0996d13ae7a" exitCode=0 Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.145384 4767 generic.go:334] "Generic (PLEG): container finished" podID="32c935bf-9481-4636-802c-3ad3b9082531" containerID="089c8fbd067c8329f240985ab1e4b647aafe09a409bbf0b7b9643b3e3dea65c1" exitCode=143 Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.145469 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"32c935bf-9481-4636-802c-3ad3b9082531","Type":"ContainerDied","Data":"40df73eddb1de8a40bda91f5552300a5e2d28eb41e9f7230bbc8d0996d13ae7a"} Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.145516 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"32c935bf-9481-4636-802c-3ad3b9082531","Type":"ContainerDied","Data":"089c8fbd067c8329f240985ab1e4b647aafe09a409bbf0b7b9643b3e3dea65c1"} Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.155082 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3c2b25b-eec8-4883-80de-f10237d9697d","Type":"ContainerDied","Data":"39929050b6cd1c12eab8ac758fb9e75e46b337f0eb9a904d94bf7188aa637b5d"} Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.155146 4767 scope.go:117] "RemoveContainer" containerID="574213173cd2f9deec1dca3468ae8fac70759231f0f10274544925ad9f23671c" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.155197 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.212398 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.240025 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.256036 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:02:32 crc kubenswrapper[4767]: E0317 16:02:32.257001 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3c2b25b-eec8-4883-80de-f10237d9697d" containerName="glance-httpd" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.257029 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3c2b25b-eec8-4883-80de-f10237d9697d" containerName="glance-httpd" Mar 17 16:02:32 crc kubenswrapper[4767]: E0317 16:02:32.257048 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3c2b25b-eec8-4883-80de-f10237d9697d" containerName="glance-log" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.257060 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3c2b25b-eec8-4883-80de-f10237d9697d" containerName="glance-log" Mar 17 16:02:32 crc kubenswrapper[4767]: E0317 16:02:32.257080 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24b48278-cd39-4c5e-b2e2-97ef9f609fa0" containerName="init" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.257088 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="24b48278-cd39-4c5e-b2e2-97ef9f609fa0" containerName="init" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.257524 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="24b48278-cd39-4c5e-b2e2-97ef9f609fa0" containerName="init" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.257559 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3c2b25b-eec8-4883-80de-f10237d9697d" containerName="glance-httpd" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.257573 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3c2b25b-eec8-4883-80de-f10237d9697d" containerName="glance-log" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.259254 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.265829 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.266261 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.298390 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.443117 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-logs\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.443257 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.443633 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9mp9\" (UniqueName: \"kubernetes.io/projected/302b036a-e22c-4c7d-a924-aa2aa150f81a-kube-api-access-z9mp9\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.444741 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.444918 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.445149 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.445201 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-scripts\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.445277 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-config-data\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.550600 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-config-data\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.550981 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-logs\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.551079 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.551319 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9mp9\" (UniqueName: \"kubernetes.io/projected/302b036a-e22c-4c7d-a924-aa2aa150f81a-kube-api-access-z9mp9\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.551404 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.551568 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.551689 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-scripts\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.551791 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.552387 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.553738 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-logs\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.560940 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-scripts\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.562423 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.562467 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/683557e62de6f02370abe27ab62c9a425f8a422dda40ecee534d661837e8035d/globalmount\"" pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.566558 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-config-data\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.567941 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.568551 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.584358 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9mp9\" (UniqueName: \"kubernetes.io/projected/302b036a-e22c-4c7d-a924-aa2aa150f81a-kube-api-access-z9mp9\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.623781 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " pod="openstack/glance-default-external-api-0" Mar 17 16:02:32 crc kubenswrapper[4767]: I0317 16:02:32.884085 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 16:02:33 crc kubenswrapper[4767]: I0317 16:02:33.297527 4767 generic.go:334] "Generic (PLEG): container finished" podID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" containerID="65995c95f1119cef59d260897bd801a022b8dec2d4889addcfd1373665067005" exitCode=0 Mar 17 16:02:33 crc kubenswrapper[4767]: I0317 16:02:33.297607 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" event={"ID":"7d4fdc1a-5119-445d-82fd-7f25fd5c879b","Type":"ContainerDied","Data":"65995c95f1119cef59d260897bd801a022b8dec2d4889addcfd1373665067005"} Mar 17 16:02:33 crc kubenswrapper[4767]: I0317 16:02:33.299113 4767 generic.go:334] "Generic (PLEG): container finished" podID="4ee82178-eb34-4b0e-9fa1-ffa7b786431d" containerID="0dab081d8810ac517e5792d0ceeee372d5ab8873ed4c7434b2d8caee8d2deda1" exitCode=0 Mar 17 16:02:33 crc kubenswrapper[4767]: I0317 16:02:33.299199 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fxjjm" event={"ID":"4ee82178-eb34-4b0e-9fa1-ffa7b786431d","Type":"ContainerDied","Data":"0dab081d8810ac517e5792d0ceeee372d5ab8873ed4c7434b2d8caee8d2deda1"} Mar 17 16:02:33 crc kubenswrapper[4767]: I0317 16:02:33.389879 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3c2b25b-eec8-4883-80de-f10237d9697d" path="/var/lib/kubelet/pods/a3c2b25b-eec8-4883-80de-f10237d9697d/volumes" Mar 17 16:02:33 crc kubenswrapper[4767]: I0317 16:02:33.557933 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" podUID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.174:5353: connect: connection refused" Mar 17 16:02:34 crc kubenswrapper[4767]: I0317 16:02:34.455022 4767 scope.go:117] "RemoveContainer" containerID="d15a14c592e76d72a87f044c7fa3daad5448e5a593f3caa5fd73086a616b8c52" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.066162 4767 scope.go:117] "RemoveContainer" containerID="5632c8f1a2ead0b8063f52708160719115c6d7fc19cc764d22f41e8d1708f1f3" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.263500 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.274522 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.373329 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95z5m\" (UniqueName: \"kubernetes.io/projected/32c935bf-9481-4636-802c-3ad3b9082531-kube-api-access-95z5m\") pod \"32c935bf-9481-4636-802c-3ad3b9082531\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.373428 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-httpd-run\") pod \"32c935bf-9481-4636-802c-3ad3b9082531\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.373486 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-logs\") pod \"32c935bf-9481-4636-802c-3ad3b9082531\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.373572 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-scripts\") pod \"32c935bf-9481-4636-802c-3ad3b9082531\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.373855 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"32c935bf-9481-4636-802c-3ad3b9082531\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.373903 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-internal-tls-certs\") pod \"32c935bf-9481-4636-802c-3ad3b9082531\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.373926 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-combined-ca-bundle\") pod \"32c935bf-9481-4636-802c-3ad3b9082531\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.373993 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-config-data\") pod \"32c935bf-9481-4636-802c-3ad3b9082531\" (UID: \"32c935bf-9481-4636-802c-3ad3b9082531\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.377815 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-logs" (OuterVolumeSpecName: "logs") pod "32c935bf-9481-4636-802c-3ad3b9082531" (UID: "32c935bf-9481-4636-802c-3ad3b9082531"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.378410 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "32c935bf-9481-4636-802c-3ad3b9082531" (UID: "32c935bf-9481-4636-802c-3ad3b9082531"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.383303 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-scripts" (OuterVolumeSpecName: "scripts") pod "32c935bf-9481-4636-802c-3ad3b9082531" (UID: "32c935bf-9481-4636-802c-3ad3b9082531"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.386975 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32c935bf-9481-4636-802c-3ad3b9082531-kube-api-access-95z5m" (OuterVolumeSpecName: "kube-api-access-95z5m") pod "32c935bf-9481-4636-802c-3ad3b9082531" (UID: "32c935bf-9481-4636-802c-3ad3b9082531"). InnerVolumeSpecName "kube-api-access-95z5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.810217 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-config-data\") pod \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.810294 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-combined-ca-bundle\") pod \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.810335 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7dpc\" (UniqueName: \"kubernetes.io/projected/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-kube-api-access-f7dpc\") pod \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.810374 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-scripts\") pod \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.810400 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-fernet-keys\") pod \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.821611 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-credential-keys\") pod \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\" (UID: \"4ee82178-eb34-4b0e-9fa1-ffa7b786431d\") " Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.843595 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95z5m\" (UniqueName: \"kubernetes.io/projected/32c935bf-9481-4636-802c-3ad3b9082531-kube-api-access-95z5m\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.843856 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4ee82178-eb34-4b0e-9fa1-ffa7b786431d" (UID: "4ee82178-eb34-4b0e-9fa1-ffa7b786431d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.843911 4767 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.843927 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32c935bf-9481-4636-802c-3ad3b9082531-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.843936 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.857667 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4ee82178-eb34-4b0e-9fa1-ffa7b786431d" (UID: "4ee82178-eb34-4b0e-9fa1-ffa7b786431d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.857934 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-kube-api-access-f7dpc" (OuterVolumeSpecName: "kube-api-access-f7dpc") pod "4ee82178-eb34-4b0e-9fa1-ffa7b786431d" (UID: "4ee82178-eb34-4b0e-9fa1-ffa7b786431d"). InnerVolumeSpecName "kube-api-access-f7dpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.896462 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-scripts" (OuterVolumeSpecName: "scripts") pod "4ee82178-eb34-4b0e-9fa1-ffa7b786431d" (UID: "4ee82178-eb34-4b0e-9fa1-ffa7b786431d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.898823 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"32c935bf-9481-4636-802c-3ad3b9082531","Type":"ContainerDied","Data":"c1c832781d89720410ae811bef7951cb32893c83ca15cde8e83a46d77e68b338"} Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.898970 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.904779 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fxjjm" event={"ID":"4ee82178-eb34-4b0e-9fa1-ffa7b786431d","Type":"ContainerDied","Data":"3609b35d210f32bce8b66bb1b7cb0f921b1e89f1b0a7a7a9997cc0911c5f8607"} Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.904836 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3609b35d210f32bce8b66bb1b7cb0f921b1e89f1b0a7a7a9997cc0911c5f8607" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.904930 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fxjjm" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.950311 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fxjjm"] Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.957151 4767 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.957207 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7dpc\" (UniqueName: \"kubernetes.io/projected/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-kube-api-access-f7dpc\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.957221 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.957232 4767 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:35 crc kubenswrapper[4767]: I0317 16:02:35.975402 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32c935bf-9481-4636-802c-3ad3b9082531" (UID: "32c935bf-9481-4636-802c-3ad3b9082531"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.040026 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fxjjm"] Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.067864 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581" (OuterVolumeSpecName: "glance") pod "32c935bf-9481-4636-802c-3ad3b9082531" (UID: "32c935bf-9481-4636-802c-3ad3b9082531"). InnerVolumeSpecName "pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.089569 4767 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") on node \"crc\" " Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.089613 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.105179 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "32c935bf-9481-4636-802c-3ad3b9082531" (UID: "32c935bf-9481-4636-802c-3ad3b9082531"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.111047 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ee82178-eb34-4b0e-9fa1-ffa7b786431d" (UID: "4ee82178-eb34-4b0e-9fa1-ffa7b786431d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.111658 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-vhwq7"] Mar 17 16:02:36 crc kubenswrapper[4767]: E0317 16:02:36.112572 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c935bf-9481-4636-802c-3ad3b9082531" containerName="glance-httpd" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.112609 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c935bf-9481-4636-802c-3ad3b9082531" containerName="glance-httpd" Mar 17 16:02:36 crc kubenswrapper[4767]: E0317 16:02:36.112662 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c935bf-9481-4636-802c-3ad3b9082531" containerName="glance-log" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.112673 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c935bf-9481-4636-802c-3ad3b9082531" containerName="glance-log" Mar 17 16:02:36 crc kubenswrapper[4767]: E0317 16:02:36.112697 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ee82178-eb34-4b0e-9fa1-ffa7b786431d" containerName="keystone-bootstrap" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.112708 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ee82178-eb34-4b0e-9fa1-ffa7b786431d" containerName="keystone-bootstrap" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.112986 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ee82178-eb34-4b0e-9fa1-ffa7b786431d" containerName="keystone-bootstrap" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.113013 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c935bf-9481-4636-802c-3ad3b9082531" containerName="glance-httpd" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.113036 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c935bf-9481-4636-802c-3ad3b9082531" containerName="glance-log" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.116844 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.157094 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-config-data" (OuterVolumeSpecName: "config-data") pod "4ee82178-eb34-4b0e-9fa1-ffa7b786431d" (UID: "4ee82178-eb34-4b0e-9fa1-ffa7b786431d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.158018 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-config-data" (OuterVolumeSpecName: "config-data") pod "32c935bf-9481-4636-802c-3ad3b9082531" (UID: "32c935bf-9481-4636-802c-3ad3b9082531"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.165644 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vhwq7"] Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.197049 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-fernet-keys\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.197253 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-scripts\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.197314 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9clmn\" (UniqueName: \"kubernetes.io/projected/f2af7040-49bc-4689-8af6-a7e3a37db7f2-kube-api-access-9clmn\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.197585 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-combined-ca-bundle\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.197700 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-config-data\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.197855 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-credential-keys\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.198061 4767 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.198088 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32c935bf-9481-4636-802c-3ad3b9082531-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.198101 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.198113 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee82178-eb34-4b0e-9fa1-ffa7b786431d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.202092 4767 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.202391 4767 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581") on node "crc" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.301923 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-fernet-keys\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.302080 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-scripts\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.302137 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9clmn\" (UniqueName: \"kubernetes.io/projected/f2af7040-49bc-4689-8af6-a7e3a37db7f2-kube-api-access-9clmn\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.302396 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-combined-ca-bundle\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.302523 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-config-data\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.302689 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-credential-keys\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.302798 4767 reconciler_common.go:293] "Volume detached for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.308665 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-scripts\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.308880 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-combined-ca-bundle\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.311836 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-fernet-keys\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.312726 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-config-data\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.313662 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.324550 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-credential-keys\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.330119 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9clmn\" (UniqueName: \"kubernetes.io/projected/f2af7040-49bc-4689-8af6-a7e3a37db7f2-kube-api-access-9clmn\") pod \"keystone-bootstrap-vhwq7\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.342984 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.378347 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.387999 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.388126 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.390767 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.391169 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.507586 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.507921 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.508385 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.508478 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.508542 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.508609 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.508829 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhlhg\" (UniqueName: \"kubernetes.io/projected/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-kube-api-access-xhlhg\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.508991 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.592053 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.611105 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.612547 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.612601 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.613840 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.613889 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.613941 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.614769 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhlhg\" (UniqueName: \"kubernetes.io/projected/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-kube-api-access-xhlhg\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.615049 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.615143 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.616289 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.616920 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.616964 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5f7c8528f2df23f78ebd936b9a4e66a32c7cd4122d28a5f082f69cd7529229c4/globalmount\"" pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.617055 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.621183 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.622785 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.629208 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.635458 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhlhg\" (UniqueName: \"kubernetes.io/projected/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-kube-api-access-xhlhg\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.671051 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:02:36 crc kubenswrapper[4767]: I0317 16:02:36.765901 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 16:02:37 crc kubenswrapper[4767]: I0317 16:02:37.374831 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32c935bf-9481-4636-802c-3ad3b9082531" path="/var/lib/kubelet/pods/32c935bf-9481-4636-802c-3ad3b9082531/volumes" Mar 17 16:02:37 crc kubenswrapper[4767]: I0317 16:02:37.376042 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ee82178-eb34-4b0e-9fa1-ffa7b786431d" path="/var/lib/kubelet/pods/4ee82178-eb34-4b0e-9fa1-ffa7b786431d/volumes" Mar 17 16:02:43 crc kubenswrapper[4767]: I0317 16:02:43.809491 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" podUID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.174:5353: i/o timeout" Mar 17 16:02:45 crc kubenswrapper[4767]: I0317 16:02:45.866315 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:45 crc kubenswrapper[4767]: I0317 16:02:45.941700 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-nb\") pod \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " Mar 17 16:02:45 crc kubenswrapper[4767]: I0317 16:02:45.941795 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-svc\") pod \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " Mar 17 16:02:45 crc kubenswrapper[4767]: I0317 16:02:45.941923 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qst6s\" (UniqueName: \"kubernetes.io/projected/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-kube-api-access-qst6s\") pod \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " Mar 17 16:02:45 crc kubenswrapper[4767]: I0317 16:02:45.941950 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-swift-storage-0\") pod \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " Mar 17 16:02:45 crc kubenswrapper[4767]: I0317 16:02:45.942011 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-sb\") pod \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " Mar 17 16:02:45 crc kubenswrapper[4767]: I0317 16:02:45.942082 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-config\") pod \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\" (UID: \"7d4fdc1a-5119-445d-82fd-7f25fd5c879b\") " Mar 17 16:02:45 crc kubenswrapper[4767]: I0317 16:02:45.995353 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-kube-api-access-qst6s" (OuterVolumeSpecName: "kube-api-access-qst6s") pod "7d4fdc1a-5119-445d-82fd-7f25fd5c879b" (UID: "7d4fdc1a-5119-445d-82fd-7f25fd5c879b"). InnerVolumeSpecName "kube-api-access-qst6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.022758 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7d4fdc1a-5119-445d-82fd-7f25fd5c879b" (UID: "7d4fdc1a-5119-445d-82fd-7f25fd5c879b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.027699 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-config" (OuterVolumeSpecName: "config") pod "7d4fdc1a-5119-445d-82fd-7f25fd5c879b" (UID: "7d4fdc1a-5119-445d-82fd-7f25fd5c879b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.036379 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7d4fdc1a-5119-445d-82fd-7f25fd5c879b" (UID: "7d4fdc1a-5119-445d-82fd-7f25fd5c879b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.048683 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.048724 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qst6s\" (UniqueName: \"kubernetes.io/projected/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-kube-api-access-qst6s\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.048741 4767 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.048752 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.049364 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7d4fdc1a-5119-445d-82fd-7f25fd5c879b" (UID: "7d4fdc1a-5119-445d-82fd-7f25fd5c879b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.059920 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" event={"ID":"7d4fdc1a-5119-445d-82fd-7f25fd5c879b","Type":"ContainerDied","Data":"be7af1f32b1e05cebe1c09d9405387a59e0f5af9b49c19244bdda3043ea1f85d"} Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.060041 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.062707 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7d4fdc1a-5119-445d-82fd-7f25fd5c879b" (UID: "7d4fdc1a-5119-445d-82fd-7f25fd5c879b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.151346 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.151393 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d4fdc1a-5119-445d-82fd-7f25fd5c879b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.390332 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.401045 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-c7c7p"] Mar 17 16:02:46 crc kubenswrapper[4767]: I0317 16:02:46.413107 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-c7c7p"] Mar 17 16:02:47 crc kubenswrapper[4767]: E0317 16:02:47.309029 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Mar 17 16:02:47 crc kubenswrapper[4767]: E0317 16:02:47.309350 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5glpn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-xj9pj_openstack(850bce7d-74fa-4ead-b6c2-fcf21c3083ec): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:02:47 crc kubenswrapper[4767]: E0317 16:02:47.310556 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-xj9pj" podUID="850bce7d-74fa-4ead-b6c2-fcf21c3083ec" Mar 17 16:02:47 crc kubenswrapper[4767]: I0317 16:02:47.368944 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" path="/var/lib/kubelet/pods/7d4fdc1a-5119-445d-82fd-7f25fd5c879b/volumes" Mar 17 16:02:48 crc kubenswrapper[4767]: E0317 16:02:48.289896 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-xj9pj" podUID="850bce7d-74fa-4ead-b6c2-fcf21c3083ec" Mar 17 16:02:48 crc kubenswrapper[4767]: I0317 16:02:48.814600 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-c7c7p" podUID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.174:5353: i/o timeout" Mar 17 16:02:50 crc kubenswrapper[4767]: I0317 16:02:50.775770 4767 generic.go:334] "Generic (PLEG): container finished" podID="958adc8c-3c1c-4b35-8cc0-fa7a3b600842" containerID="207a85d187102e1a1442134ad96bbcd58337d2da80439bd774ba9ac839b6831e" exitCode=0 Mar 17 16:02:50 crc kubenswrapper[4767]: I0317 16:02:50.776348 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gm7c4" event={"ID":"958adc8c-3c1c-4b35-8cc0-fa7a3b600842","Type":"ContainerDied","Data":"207a85d187102e1a1442134ad96bbcd58337d2da80439bd774ba9ac839b6831e"} Mar 17 16:02:59 crc kubenswrapper[4767]: E0317 16:02:59.036972 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Mar 17 16:02:59 crc kubenswrapper[4767]: E0317 16:02:59.037894 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z8kxb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-657sp_openstack(5e6a688d-7438-45a1-9aa7-b7042485163f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:02:59 crc kubenswrapper[4767]: E0317 16:02:59.039260 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-657sp" podUID="5e6a688d-7438-45a1-9aa7-b7042485163f" Mar 17 16:02:59 crc kubenswrapper[4767]: E0317 16:02:59.841937 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Mar 17 16:02:59 crc kubenswrapper[4767]: E0317 16:02:59.842365 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n94h5bh65ch5bbhbch575h5bh546hch5dfh8h57dh586h9ch57ch555hd7h65bh685h557h57ch77hc9h8dh5c9h685h5d9h64ch54h566h5cfh55cq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-25njk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(d17172fd-996e-496e-92b5-9cd19edc65f4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:02:59 crc kubenswrapper[4767]: I0317 16:02:59.876317 4767 scope.go:117] "RemoveContainer" containerID="40df73eddb1de8a40bda91f5552300a5e2d28eb41e9f7230bbc8d0996d13ae7a" Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.008056 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gm7c4" event={"ID":"958adc8c-3c1c-4b35-8cc0-fa7a3b600842","Type":"ContainerDied","Data":"e4ce5c7ada7a6cfb87e8eeae2a4274ae320237b8835e7247161ddc3add750393"} Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.008427 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4ce5c7ada7a6cfb87e8eeae2a4274ae320237b8835e7247161ddc3add750393" Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.010845 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"302b036a-e22c-4c7d-a924-aa2aa150f81a","Type":"ContainerStarted","Data":"60742eebfd988bd712d7960f07123e36bf490d29d5121dd9af00263459595593"} Mar 17 16:03:00 crc kubenswrapper[4767]: E0317 16:03:00.014094 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-657sp" podUID="5e6a688d-7438-45a1-9aa7-b7042485163f" Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.104425 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.150429 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-combined-ca-bundle\") pod \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.150588 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz6rx\" (UniqueName: \"kubernetes.io/projected/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-kube-api-access-zz6rx\") pod \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.150652 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-config\") pod \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\" (UID: \"958adc8c-3c1c-4b35-8cc0-fa7a3b600842\") " Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.160772 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-kube-api-access-zz6rx" (OuterVolumeSpecName: "kube-api-access-zz6rx") pod "958adc8c-3c1c-4b35-8cc0-fa7a3b600842" (UID: "958adc8c-3c1c-4b35-8cc0-fa7a3b600842"). InnerVolumeSpecName "kube-api-access-zz6rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.179979 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "958adc8c-3c1c-4b35-8cc0-fa7a3b600842" (UID: "958adc8c-3c1c-4b35-8cc0-fa7a3b600842"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.188485 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-config" (OuterVolumeSpecName: "config") pod "958adc8c-3c1c-4b35-8cc0-fa7a3b600842" (UID: "958adc8c-3c1c-4b35-8cc0-fa7a3b600842"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.262494 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.262815 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz6rx\" (UniqueName: \"kubernetes.io/projected/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-kube-api-access-zz6rx\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:00 crc kubenswrapper[4767]: I0317 16:03:00.262828 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/958adc8c-3c1c-4b35-8cc0-fa7a3b600842-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.020960 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gm7c4" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.319869 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kh22t"] Mar 17 16:03:01 crc kubenswrapper[4767]: E0317 16:03:01.321328 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" containerName="init" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.321359 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" containerName="init" Mar 17 16:03:01 crc kubenswrapper[4767]: E0317 16:03:01.321405 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="958adc8c-3c1c-4b35-8cc0-fa7a3b600842" containerName="neutron-db-sync" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.321415 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="958adc8c-3c1c-4b35-8cc0-fa7a3b600842" containerName="neutron-db-sync" Mar 17 16:03:01 crc kubenswrapper[4767]: E0317 16:03:01.321492 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" containerName="dnsmasq-dns" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.321506 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" containerName="dnsmasq-dns" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.322048 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d4fdc1a-5119-445d-82fd-7f25fd5c879b" containerName="dnsmasq-dns" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.322092 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="958adc8c-3c1c-4b35-8cc0-fa7a3b600842" containerName="neutron-db-sync" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.324080 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.344509 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kh22t"] Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.455854 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-667f69cbdb-5qvs7"] Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.459890 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.469108 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.469984 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-2gvvz" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.473418 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.473668 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.475609 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84kkr\" (UniqueName: \"kubernetes.io/projected/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-kube-api-access-84kkr\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.475702 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-config\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.475732 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-svc\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.475763 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.475860 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.475986 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.478719 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-667f69cbdb-5qvs7"] Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.580913 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.578883 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.581053 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84kkr\" (UniqueName: \"kubernetes.io/projected/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-kube-api-access-84kkr\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.581137 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rlgp\" (UniqueName: \"kubernetes.io/projected/86426b6e-ec55-49d4-804e-75b204239fa1-kube-api-access-8rlgp\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.581213 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-config\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.581238 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-config\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.581277 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-combined-ca-bundle\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.581302 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-svc\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.581367 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.581518 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.584030 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-ovndb-tls-certs\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.584125 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-httpd-config\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.583781 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.582713 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-config\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.583150 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-svc\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.584613 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.608382 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84kkr\" (UniqueName: \"kubernetes.io/projected/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-kube-api-access-84kkr\") pod \"dnsmasq-dns-55f844cf75-kh22t\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.688659 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rlgp\" (UniqueName: \"kubernetes.io/projected/86426b6e-ec55-49d4-804e-75b204239fa1-kube-api-access-8rlgp\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.688835 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-config\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.688900 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-combined-ca-bundle\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.689095 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-ovndb-tls-certs\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.691614 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-httpd-config\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.695352 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-config\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.695936 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-httpd-config\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.697187 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-ovndb-tls-certs\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.707749 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.709621 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-combined-ca-bundle\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.713789 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rlgp\" (UniqueName: \"kubernetes.io/projected/86426b6e-ec55-49d4-804e-75b204239fa1-kube-api-access-8rlgp\") pod \"neutron-667f69cbdb-5qvs7\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:01 crc kubenswrapper[4767]: I0317 16:03:01.831157 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:04 crc kubenswrapper[4767]: I0317 16:03:04.239442 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:03:04 crc kubenswrapper[4767]: I0317 16:03:04.239777 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.837327 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-67c55f74cc-tpfc4"] Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.840870 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.843231 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.843981 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.862075 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67c55f74cc-tpfc4"] Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.979416 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-internal-tls-certs\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.979616 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-public-tls-certs\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.979656 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-httpd-config\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.979746 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-combined-ca-bundle\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.979778 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-ovndb-tls-certs\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.979834 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2kzk\" (UniqueName: \"kubernetes.io/projected/16264c79-b41d-43e1-a692-e084ae52e928-kube-api-access-s2kzk\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:05 crc kubenswrapper[4767]: I0317 16:03:05.979880 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-config\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.082060 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-combined-ca-bundle\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.083486 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-ovndb-tls-certs\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.083977 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kzk\" (UniqueName: \"kubernetes.io/projected/16264c79-b41d-43e1-a692-e084ae52e928-kube-api-access-s2kzk\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.084094 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-config\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.085842 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-internal-tls-certs\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.086153 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-public-tls-certs\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.086278 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-httpd-config\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.090201 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-internal-tls-certs\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.090549 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-ovndb-tls-certs\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.091524 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-combined-ca-bundle\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.093291 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-public-tls-certs\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.095977 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-httpd-config\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.098149 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-config\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.115246 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kzk\" (UniqueName: \"kubernetes.io/projected/16264c79-b41d-43e1-a692-e084ae52e928-kube-api-access-s2kzk\") pod \"neutron-67c55f74cc-tpfc4\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:06 crc kubenswrapper[4767]: I0317 16:03:06.395293 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:07 crc kubenswrapper[4767]: I0317 16:03:07.617857 4767 scope.go:117] "RemoveContainer" containerID="089c8fbd067c8329f240985ab1e4b647aafe09a409bbf0b7b9643b3e3dea65c1" Mar 17 16:03:07 crc kubenswrapper[4767]: E0317 16:03:07.658254 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Mar 17 16:03:07 crc kubenswrapper[4767]: E0317 16:03:07.658481 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6jmsj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-pdxns_openstack(712803ce-6e47-40a6-bf5c-84e8aea748e4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:03:07 crc kubenswrapper[4767]: E0317 16:03:07.659795 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-pdxns" podUID="712803ce-6e47-40a6-bf5c-84e8aea748e4" Mar 17 16:03:07 crc kubenswrapper[4767]: I0317 16:03:07.819126 4767 scope.go:117] "RemoveContainer" containerID="65995c95f1119cef59d260897bd801a022b8dec2d4889addcfd1373665067005" Mar 17 16:03:08 crc kubenswrapper[4767]: I0317 16:03:08.280335 4767 scope.go:117] "RemoveContainer" containerID="a60de704a8164f69ec89d8cb9100672d84bc03253f0d8eed00f04ef221a1ec35" Mar 17 16:03:08 crc kubenswrapper[4767]: W0317 16:03:08.454017 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2af7040_49bc_4689_8af6_a7e3a37db7f2.slice/crio-3fafe533bdf56ba29ca8bba20c1d4042390d6946eafae3d9538fe3034a5a1ef3 WatchSource:0}: Error finding container 3fafe533bdf56ba29ca8bba20c1d4042390d6946eafae3d9538fe3034a5a1ef3: Status 404 returned error can't find the container with id 3fafe533bdf56ba29ca8bba20c1d4042390d6946eafae3d9538fe3034a5a1ef3 Mar 17 16:03:08 crc kubenswrapper[4767]: I0317 16:03:08.483222 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vhwq7"] Mar 17 16:03:08 crc kubenswrapper[4767]: I0317 16:03:08.903057 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:03:08 crc kubenswrapper[4767]: I0317 16:03:08.923645 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vhwq7" event={"ID":"f2af7040-49bc-4689-8af6-a7e3a37db7f2","Type":"ContainerStarted","Data":"3fafe533bdf56ba29ca8bba20c1d4042390d6946eafae3d9538fe3034a5a1ef3"} Mar 17 16:03:08 crc kubenswrapper[4767]: W0317 16:03:08.925106 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a02e83b_1e68_446f_9fd7_9cb6fb125a6f.slice/crio-24af3b22057bd463148d003486e027125a2c4b49bf6932f4acd657274a8cdbdf WatchSource:0}: Error finding container 24af3b22057bd463148d003486e027125a2c4b49bf6932f4acd657274a8cdbdf: Status 404 returned error can't find the container with id 24af3b22057bd463148d003486e027125a2c4b49bf6932f4acd657274a8cdbdf Mar 17 16:03:08 crc kubenswrapper[4767]: I0317 16:03:08.942808 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mzst5" event={"ID":"c5b58621-27a1-479d-a90c-aa96400e1116","Type":"ContainerStarted","Data":"4b19420f7562b7140f30f8aa7b439c3e38ff297bd754e956e70d52d1805782e0"} Mar 17 16:03:09 crc kubenswrapper[4767]: I0317 16:03:09.005143 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kh22t"] Mar 17 16:03:09 crc kubenswrapper[4767]: I0317 16:03:09.016969 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-mzst5" podStartSLOduration=11.986558752 podStartE2EDuration="48.016919897s" podCreationTimestamp="2026-03-17 16:02:21 +0000 UTC" firstStartedPulling="2026-03-17 16:02:23.881313996 +0000 UTC m=+1535.294630043" lastFinishedPulling="2026-03-17 16:02:59.911675141 +0000 UTC m=+1571.324991188" observedRunningTime="2026-03-17 16:03:08.970718743 +0000 UTC m=+1580.384034810" watchObservedRunningTime="2026-03-17 16:03:09.016919897 +0000 UTC m=+1580.430235944" Mar 17 16:03:09 crc kubenswrapper[4767]: E0317 16:03:09.065115 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-pdxns" podUID="712803ce-6e47-40a6-bf5c-84e8aea748e4" Mar 17 16:03:09 crc kubenswrapper[4767]: I0317 16:03:09.267129 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67c55f74cc-tpfc4"] Mar 17 16:03:10 crc kubenswrapper[4767]: I0317 16:03:10.054371 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"302b036a-e22c-4c7d-a924-aa2aa150f81a","Type":"ContainerStarted","Data":"f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156"} Mar 17 16:03:10 crc kubenswrapper[4767]: I0317 16:03:10.057248 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c55f74cc-tpfc4" event={"ID":"16264c79-b41d-43e1-a692-e084ae52e928","Type":"ContainerStarted","Data":"2f1b9b811014cbdaba721aa6146c407c4153b6cf9076e2b9678adc5f0af21d0f"} Mar 17 16:03:10 crc kubenswrapper[4767]: I0317 16:03:10.073087 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xj9pj" event={"ID":"850bce7d-74fa-4ead-b6c2-fcf21c3083ec","Type":"ContainerStarted","Data":"d83d0df8dc6726bc21c1a068465bd820f812594d7608c6a517469016f7b477dd"} Mar 17 16:03:10 crc kubenswrapper[4767]: I0317 16:03:10.088725 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vhwq7" event={"ID":"f2af7040-49bc-4689-8af6-a7e3a37db7f2","Type":"ContainerStarted","Data":"2fb9a9b45b63e921103b8b488be5012d01d9d2120233755ddea3cbee0a4f05b4"} Mar 17 16:03:10 crc kubenswrapper[4767]: I0317 16:03:10.092526 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f","Type":"ContainerStarted","Data":"24af3b22057bd463148d003486e027125a2c4b49bf6932f4acd657274a8cdbdf"} Mar 17 16:03:10 crc kubenswrapper[4767]: I0317 16:03:10.115440 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" event={"ID":"b7d5a577-7603-4cfe-a0c9-a899eeaab93d","Type":"ContainerStarted","Data":"bd16df8e254f555876f90d11edba095368d04cffdb39159ec0dde01b5e308819"} Mar 17 16:03:10 crc kubenswrapper[4767]: I0317 16:03:10.115885 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" event={"ID":"b7d5a577-7603-4cfe-a0c9-a899eeaab93d","Type":"ContainerStarted","Data":"48421ec571d9b3ae98f80a88d66d8c7f416a09279d4899bd338c256b2cd6832e"} Mar 17 16:03:10 crc kubenswrapper[4767]: I0317 16:03:10.144712 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-xj9pj" podStartSLOduration=5.829995004 podStartE2EDuration="50.144676127s" podCreationTimestamp="2026-03-17 16:02:20 +0000 UTC" firstStartedPulling="2026-03-17 16:02:23.980847324 +0000 UTC m=+1535.394163371" lastFinishedPulling="2026-03-17 16:03:08.295528447 +0000 UTC m=+1579.708844494" observedRunningTime="2026-03-17 16:03:10.10285502 +0000 UTC m=+1581.516171057" watchObservedRunningTime="2026-03-17 16:03:10.144676127 +0000 UTC m=+1581.557992174" Mar 17 16:03:10 crc kubenswrapper[4767]: I0317 16:03:10.172242 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-667f69cbdb-5qvs7"] Mar 17 16:03:10 crc kubenswrapper[4767]: I0317 16:03:10.187357 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-vhwq7" podStartSLOduration=35.187323065 podStartE2EDuration="35.187323065s" podCreationTimestamp="2026-03-17 16:02:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:10.128969977 +0000 UTC m=+1581.542286034" watchObservedRunningTime="2026-03-17 16:03:10.187323065 +0000 UTC m=+1581.600639132" Mar 17 16:03:10 crc kubenswrapper[4767]: W0317 16:03:10.466051 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86426b6e_ec55_49d4_804e_75b204239fa1.slice/crio-12eeed4dfedd2c540c520f5721a3b510295e1093b21b2b1ca3a1610081bf9ad0 WatchSource:0}: Error finding container 12eeed4dfedd2c540c520f5721a3b510295e1093b21b2b1ca3a1610081bf9ad0: Status 404 returned error can't find the container with id 12eeed4dfedd2c540c520f5721a3b510295e1093b21b2b1ca3a1610081bf9ad0 Mar 17 16:03:11 crc kubenswrapper[4767]: I0317 16:03:11.159152 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-667f69cbdb-5qvs7" event={"ID":"86426b6e-ec55-49d4-804e-75b204239fa1","Type":"ContainerStarted","Data":"12eeed4dfedd2c540c520f5721a3b510295e1093b21b2b1ca3a1610081bf9ad0"} Mar 17 16:03:11 crc kubenswrapper[4767]: I0317 16:03:11.179767 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f","Type":"ContainerStarted","Data":"35c908e5327b918655403169d8b84c8c61ba41ef35f65ece1a57a0e8e323fbab"} Mar 17 16:03:11 crc kubenswrapper[4767]: I0317 16:03:11.188228 4767 generic.go:334] "Generic (PLEG): container finished" podID="b7d5a577-7603-4cfe-a0c9-a899eeaab93d" containerID="bd16df8e254f555876f90d11edba095368d04cffdb39159ec0dde01b5e308819" exitCode=0 Mar 17 16:03:11 crc kubenswrapper[4767]: I0317 16:03:11.189409 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" event={"ID":"b7d5a577-7603-4cfe-a0c9-a899eeaab93d","Type":"ContainerDied","Data":"bd16df8e254f555876f90d11edba095368d04cffdb39159ec0dde01b5e308819"} Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.207733 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" event={"ID":"b7d5a577-7603-4cfe-a0c9-a899eeaab93d","Type":"ContainerStarted","Data":"0f423a7202aa9a2dde93bb3217ea9dd20ec328c78d33d70274b1ec90128c1472"} Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.208504 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.210865 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"302b036a-e22c-4c7d-a924-aa2aa150f81a","Type":"ContainerStarted","Data":"02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e"} Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.214938 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c55f74cc-tpfc4" event={"ID":"16264c79-b41d-43e1-a692-e084ae52e928","Type":"ContainerStarted","Data":"359eb33421735c4d333f5eede05ad9d61866eb0e09e62023bb64790d496a4e82"} Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.214999 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c55f74cc-tpfc4" event={"ID":"16264c79-b41d-43e1-a692-e084ae52e928","Type":"ContainerStarted","Data":"0806cabe6f745515e02e3ee6523619eef04a5837a0805c88be28d9b6873fca60"} Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.215109 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.220087 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-667f69cbdb-5qvs7" event={"ID":"86426b6e-ec55-49d4-804e-75b204239fa1","Type":"ContainerStarted","Data":"d8b69c7107b0c5f91af953d758844dc3ee89532cd4ca4ee00b6b4829ad5f3f18"} Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.220138 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-667f69cbdb-5qvs7" event={"ID":"86426b6e-ec55-49d4-804e-75b204239fa1","Type":"ContainerStarted","Data":"c2b9c871dd75268d8313f0a8947b962e08fc00dd9b8852847187eaeebe41cf7d"} Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.220454 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.223901 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f","Type":"ContainerStarted","Data":"a593e4fd0fc22e116160db18ca3009f07986bca775a60abd49ebe1115947d2bf"} Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.228300 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17172fd-996e-496e-92b5-9cd19edc65f4","Type":"ContainerStarted","Data":"94848df684c7d16e35f3425fecf44bad6bb45f6cba749b60673a0a0b7fe71222"} Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.639980 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" podStartSLOduration=11.639948006000001 podStartE2EDuration="11.639948006s" podCreationTimestamp="2026-03-17 16:03:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:12.249387129 +0000 UTC m=+1583.662703186" watchObservedRunningTime="2026-03-17 16:03:12.639948006 +0000 UTC m=+1584.053264053" Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.692058 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-667f69cbdb-5qvs7" podStartSLOduration=11.692029395 podStartE2EDuration="11.692029395s" podCreationTimestamp="2026-03-17 16:03:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:12.677744047 +0000 UTC m=+1584.091060094" watchObservedRunningTime="2026-03-17 16:03:12.692029395 +0000 UTC m=+1584.105345442" Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.741627 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=40.741594133 podStartE2EDuration="40.741594133s" podCreationTimestamp="2026-03-17 16:02:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:12.739843573 +0000 UTC m=+1584.153159630" watchObservedRunningTime="2026-03-17 16:03:12.741594133 +0000 UTC m=+1584.154910180" Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.770906 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=36.77087613 podStartE2EDuration="36.77087613s" podCreationTimestamp="2026-03-17 16:02:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:12.764124017 +0000 UTC m=+1584.177440074" watchObservedRunningTime="2026-03-17 16:03:12.77087613 +0000 UTC m=+1584.184192177" Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.825534 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-67c55f74cc-tpfc4" podStartSLOduration=7.825504483 podStartE2EDuration="7.825504483s" podCreationTimestamp="2026-03-17 16:03:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:12.801817045 +0000 UTC m=+1584.215133092" watchObservedRunningTime="2026-03-17 16:03:12.825504483 +0000 UTC m=+1584.238820530" Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.885092 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.885154 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.947295 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 17 16:03:12 crc kubenswrapper[4767]: I0317 16:03:12.949259 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 17 16:03:13 crc kubenswrapper[4767]: I0317 16:03:13.248580 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 17 16:03:13 crc kubenswrapper[4767]: I0317 16:03:13.248619 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 17 16:03:14 crc kubenswrapper[4767]: I0317 16:03:14.263765 4767 generic.go:334] "Generic (PLEG): container finished" podID="c5b58621-27a1-479d-a90c-aa96400e1116" containerID="4b19420f7562b7140f30f8aa7b439c3e38ff297bd754e956e70d52d1805782e0" exitCode=0 Mar 17 16:03:14 crc kubenswrapper[4767]: I0317 16:03:14.263994 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mzst5" event={"ID":"c5b58621-27a1-479d-a90c-aa96400e1116","Type":"ContainerDied","Data":"4b19420f7562b7140f30f8aa7b439c3e38ff297bd754e956e70d52d1805782e0"} Mar 17 16:03:14 crc kubenswrapper[4767]: I0317 16:03:14.270795 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-657sp" event={"ID":"5e6a688d-7438-45a1-9aa7-b7042485163f","Type":"ContainerStarted","Data":"218dbac6c2516ec6291a0108ea2ab138bafea03e1445ea1e37fdbff853f3e8ac"} Mar 17 16:03:14 crc kubenswrapper[4767]: I0317 16:03:14.320859 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-657sp" podStartSLOduration=3.909013806 podStartE2EDuration="54.320828652s" podCreationTimestamp="2026-03-17 16:02:20 +0000 UTC" firstStartedPulling="2026-03-17 16:02:22.853295 +0000 UTC m=+1534.266611047" lastFinishedPulling="2026-03-17 16:03:13.265109846 +0000 UTC m=+1584.678425893" observedRunningTime="2026-03-17 16:03:14.317813876 +0000 UTC m=+1585.731130063" watchObservedRunningTime="2026-03-17 16:03:14.320828652 +0000 UTC m=+1585.734144699" Mar 17 16:03:16 crc kubenswrapper[4767]: I0317 16:03:16.538931 4767 generic.go:334] "Generic (PLEG): container finished" podID="f2af7040-49bc-4689-8af6-a7e3a37db7f2" containerID="2fb9a9b45b63e921103b8b488be5012d01d9d2120233755ddea3cbee0a4f05b4" exitCode=0 Mar 17 16:03:16 crc kubenswrapper[4767]: I0317 16:03:16.539594 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vhwq7" event={"ID":"f2af7040-49bc-4689-8af6-a7e3a37db7f2","Type":"ContainerDied","Data":"2fb9a9b45b63e921103b8b488be5012d01d9d2120233755ddea3cbee0a4f05b4"} Mar 17 16:03:16 crc kubenswrapper[4767]: I0317 16:03:16.546286 4767 generic.go:334] "Generic (PLEG): container finished" podID="850bce7d-74fa-4ead-b6c2-fcf21c3083ec" containerID="d83d0df8dc6726bc21c1a068465bd820f812594d7608c6a517469016f7b477dd" exitCode=0 Mar 17 16:03:16 crc kubenswrapper[4767]: I0317 16:03:16.546371 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xj9pj" event={"ID":"850bce7d-74fa-4ead-b6c2-fcf21c3083ec","Type":"ContainerDied","Data":"d83d0df8dc6726bc21c1a068465bd820f812594d7608c6a517469016f7b477dd"} Mar 17 16:03:16 crc kubenswrapper[4767]: I0317 16:03:16.716426 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:16 crc kubenswrapper[4767]: I0317 16:03:16.773050 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 17 16:03:16 crc kubenswrapper[4767]: I0317 16:03:16.774728 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 17 16:03:16 crc kubenswrapper[4767]: I0317 16:03:16.804015 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-h7hgh"] Mar 17 16:03:16 crc kubenswrapper[4767]: I0317 16:03:16.804355 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" podUID="4fd01229-44e5-40e9-ba3c-d672309ca045" containerName="dnsmasq-dns" containerID="cri-o://a58959ac80935fea8311b60a5217d0be5b31501cd9709bf2e61a38d6101605aa" gracePeriod=10 Mar 17 16:03:16 crc kubenswrapper[4767]: I0317 16:03:16.868539 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 17 16:03:16 crc kubenswrapper[4767]: I0317 16:03:16.890891 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 17 16:03:17 crc kubenswrapper[4767]: I0317 16:03:17.566004 4767 generic.go:334] "Generic (PLEG): container finished" podID="4fd01229-44e5-40e9-ba3c-d672309ca045" containerID="a58959ac80935fea8311b60a5217d0be5b31501cd9709bf2e61a38d6101605aa" exitCode=0 Mar 17 16:03:17 crc kubenswrapper[4767]: I0317 16:03:17.566067 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" event={"ID":"4fd01229-44e5-40e9-ba3c-d672309ca045","Type":"ContainerDied","Data":"a58959ac80935fea8311b60a5217d0be5b31501cd9709bf2e61a38d6101605aa"} Mar 17 16:03:17 crc kubenswrapper[4767]: I0317 16:03:17.566951 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 17 16:03:17 crc kubenswrapper[4767]: I0317 16:03:17.566981 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 17 16:03:20 crc kubenswrapper[4767]: I0317 16:03:20.919147 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 17 16:03:20 crc kubenswrapper[4767]: I0317 16:03:20.930378 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 17 16:03:21 crc kubenswrapper[4767]: I0317 16:03:21.281864 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 17 16:03:21 crc kubenswrapper[4767]: I0317 16:03:21.282072 4767 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 16:03:21 crc kubenswrapper[4767]: I0317 16:03:21.633987 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 17 16:03:21 crc kubenswrapper[4767]: I0317 16:03:21.660293 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" podUID="4fd01229-44e5-40e9-ba3c-d672309ca045" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.188:5353: connect: connection refused" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.382558 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.454285 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.462512 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mzst5" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.545241 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-credential-keys\") pod \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.545378 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-config-data\") pod \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.545479 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-combined-ca-bundle\") pod \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.545598 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-scripts\") pod \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.545740 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-fernet-keys\") pod \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.545767 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9clmn\" (UniqueName: \"kubernetes.io/projected/f2af7040-49bc-4689-8af6-a7e3a37db7f2-kube-api-access-9clmn\") pod \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\" (UID: \"f2af7040-49bc-4689-8af6-a7e3a37db7f2\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.597585 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f2af7040-49bc-4689-8af6-a7e3a37db7f2" (UID: "f2af7040-49bc-4689-8af6-a7e3a37db7f2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.598935 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2af7040-49bc-4689-8af6-a7e3a37db7f2-kube-api-access-9clmn" (OuterVolumeSpecName: "kube-api-access-9clmn") pod "f2af7040-49bc-4689-8af6-a7e3a37db7f2" (UID: "f2af7040-49bc-4689-8af6-a7e3a37db7f2"). InnerVolumeSpecName "kube-api-access-9clmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.616677 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-scripts" (OuterVolumeSpecName: "scripts") pod "f2af7040-49bc-4689-8af6-a7e3a37db7f2" (UID: "f2af7040-49bc-4689-8af6-a7e3a37db7f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.632541 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f2af7040-49bc-4689-8af6-a7e3a37db7f2" (UID: "f2af7040-49bc-4689-8af6-a7e3a37db7f2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.648212 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5glpn\" (UniqueName: \"kubernetes.io/projected/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-kube-api-access-5glpn\") pod \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.648320 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-db-sync-config-data\") pod \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.648351 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmtvx\" (UniqueName: \"kubernetes.io/projected/c5b58621-27a1-479d-a90c-aa96400e1116-kube-api-access-bmtvx\") pod \"c5b58621-27a1-479d-a90c-aa96400e1116\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.648376 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-scripts\") pod \"c5b58621-27a1-479d-a90c-aa96400e1116\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.648404 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-config-data\") pod \"c5b58621-27a1-479d-a90c-aa96400e1116\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.648550 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5b58621-27a1-479d-a90c-aa96400e1116-logs\") pod \"c5b58621-27a1-479d-a90c-aa96400e1116\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.648845 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-combined-ca-bundle\") pod \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\" (UID: \"850bce7d-74fa-4ead-b6c2-fcf21c3083ec\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.648891 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-combined-ca-bundle\") pod \"c5b58621-27a1-479d-a90c-aa96400e1116\" (UID: \"c5b58621-27a1-479d-a90c-aa96400e1116\") " Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.649601 4767 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.649624 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9clmn\" (UniqueName: \"kubernetes.io/projected/f2af7040-49bc-4689-8af6-a7e3a37db7f2-kube-api-access-9clmn\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.649636 4767 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.649648 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.651576 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5b58621-27a1-479d-a90c-aa96400e1116-logs" (OuterVolumeSpecName: "logs") pod "c5b58621-27a1-479d-a90c-aa96400e1116" (UID: "c5b58621-27a1-479d-a90c-aa96400e1116"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.661707 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "850bce7d-74fa-4ead-b6c2-fcf21c3083ec" (UID: "850bce7d-74fa-4ead-b6c2-fcf21c3083ec"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.669909 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xj9pj" event={"ID":"850bce7d-74fa-4ead-b6c2-fcf21c3083ec","Type":"ContainerDied","Data":"4477daea19710bf9ccf749a74478f8df1c9640d0aba65c0ad1fade1b7305c3b3"} Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.670217 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4477daea19710bf9ccf749a74478f8df1c9640d0aba65c0ad1fade1b7305c3b3" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.670326 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xj9pj" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.675409 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5b58621-27a1-479d-a90c-aa96400e1116-kube-api-access-bmtvx" (OuterVolumeSpecName: "kube-api-access-bmtvx") pod "c5b58621-27a1-479d-a90c-aa96400e1116" (UID: "c5b58621-27a1-479d-a90c-aa96400e1116"). InnerVolumeSpecName "kube-api-access-bmtvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.676101 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-scripts" (OuterVolumeSpecName: "scripts") pod "c5b58621-27a1-479d-a90c-aa96400e1116" (UID: "c5b58621-27a1-479d-a90c-aa96400e1116"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.678297 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vhwq7" event={"ID":"f2af7040-49bc-4689-8af6-a7e3a37db7f2","Type":"ContainerDied","Data":"3fafe533bdf56ba29ca8bba20c1d4042390d6946eafae3d9538fe3034a5a1ef3"} Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.678363 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fafe533bdf56ba29ca8bba20c1d4042390d6946eafae3d9538fe3034a5a1ef3" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.678470 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vhwq7" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.679608 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-kube-api-access-5glpn" (OuterVolumeSpecName: "kube-api-access-5glpn") pod "850bce7d-74fa-4ead-b6c2-fcf21c3083ec" (UID: "850bce7d-74fa-4ead-b6c2-fcf21c3083ec"). InnerVolumeSpecName "kube-api-access-5glpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.684729 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mzst5" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.684727 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mzst5" event={"ID":"c5b58621-27a1-479d-a90c-aa96400e1116","Type":"ContainerDied","Data":"9886db56504484e4251c813478a7c5d8201378f98cbc61258596d6ebd57e344b"} Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.684873 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9886db56504484e4251c813478a7c5d8201378f98cbc61258596d6ebd57e344b" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.691360 4767 generic.go:334] "Generic (PLEG): container finished" podID="5e6a688d-7438-45a1-9aa7-b7042485163f" containerID="218dbac6c2516ec6291a0108ea2ab138bafea03e1445ea1e37fdbff853f3e8ac" exitCode=0 Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.691425 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-657sp" event={"ID":"5e6a688d-7438-45a1-9aa7-b7042485163f","Type":"ContainerDied","Data":"218dbac6c2516ec6291a0108ea2ab138bafea03e1445ea1e37fdbff853f3e8ac"} Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.748342 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "850bce7d-74fa-4ead-b6c2-fcf21c3083ec" (UID: "850bce7d-74fa-4ead-b6c2-fcf21c3083ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.752043 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.752198 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5glpn\" (UniqueName: \"kubernetes.io/projected/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-kube-api-access-5glpn\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.752316 4767 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/850bce7d-74fa-4ead-b6c2-fcf21c3083ec-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.752409 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmtvx\" (UniqueName: \"kubernetes.io/projected/c5b58621-27a1-479d-a90c-aa96400e1116-kube-api-access-bmtvx\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.753326 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.753416 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5b58621-27a1-479d-a90c-aa96400e1116-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.763535 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-config-data" (OuterVolumeSpecName: "config-data") pod "c5b58621-27a1-479d-a90c-aa96400e1116" (UID: "c5b58621-27a1-479d-a90c-aa96400e1116"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.790992 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-config-data" (OuterVolumeSpecName: "config-data") pod "f2af7040-49bc-4689-8af6-a7e3a37db7f2" (UID: "f2af7040-49bc-4689-8af6-a7e3a37db7f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.804374 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2af7040-49bc-4689-8af6-a7e3a37db7f2" (UID: "f2af7040-49bc-4689-8af6-a7e3a37db7f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.821417 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5b58621-27a1-479d-a90c-aa96400e1116" (UID: "c5b58621-27a1-479d-a90c-aa96400e1116"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.855463 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.855805 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5b58621-27a1-479d-a90c-aa96400e1116-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.855894 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.855958 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2af7040-49bc-4689-8af6-a7e3a37db7f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:22 crc kubenswrapper[4767]: I0317 16:03:22.857726 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.125159 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-config\") pod \"4fd01229-44e5-40e9-ba3c-d672309ca045\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.125585 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-sb\") pod \"4fd01229-44e5-40e9-ba3c-d672309ca045\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.127272 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-swift-storage-0\") pod \"4fd01229-44e5-40e9-ba3c-d672309ca045\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.127850 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmb6t\" (UniqueName: \"kubernetes.io/projected/4fd01229-44e5-40e9-ba3c-d672309ca045-kube-api-access-hmb6t\") pod \"4fd01229-44e5-40e9-ba3c-d672309ca045\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.127919 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-nb\") pod \"4fd01229-44e5-40e9-ba3c-d672309ca045\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.128043 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-svc\") pod \"4fd01229-44e5-40e9-ba3c-d672309ca045\" (UID: \"4fd01229-44e5-40e9-ba3c-d672309ca045\") " Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.159343 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fd01229-44e5-40e9-ba3c-d672309ca045-kube-api-access-hmb6t" (OuterVolumeSpecName: "kube-api-access-hmb6t") pod "4fd01229-44e5-40e9-ba3c-d672309ca045" (UID: "4fd01229-44e5-40e9-ba3c-d672309ca045"). InnerVolumeSpecName "kube-api-access-hmb6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.186904 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4fd01229-44e5-40e9-ba3c-d672309ca045" (UID: "4fd01229-44e5-40e9-ba3c-d672309ca045"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.232126 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmb6t\" (UniqueName: \"kubernetes.io/projected/4fd01229-44e5-40e9-ba3c-d672309ca045-kube-api-access-hmb6t\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.232159 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.247370 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4fd01229-44e5-40e9-ba3c-d672309ca045" (UID: "4fd01229-44e5-40e9-ba3c-d672309ca045"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.263851 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-config" (OuterVolumeSpecName: "config") pod "4fd01229-44e5-40e9-ba3c-d672309ca045" (UID: "4fd01229-44e5-40e9-ba3c-d672309ca045"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.279740 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4fd01229-44e5-40e9-ba3c-d672309ca045" (UID: "4fd01229-44e5-40e9-ba3c-d672309ca045"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.292713 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4fd01229-44e5-40e9-ba3c-d672309ca045" (UID: "4fd01229-44e5-40e9-ba3c-d672309ca045"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.334570 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.334831 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.334979 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.335091 4767 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fd01229-44e5-40e9-ba3c-d672309ca045-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.587517 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-79f98b8574-mc4pp"] Mar 17 16:03:23 crc kubenswrapper[4767]: E0317 16:03:23.597032 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5b58621-27a1-479d-a90c-aa96400e1116" containerName="placement-db-sync" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.597083 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5b58621-27a1-479d-a90c-aa96400e1116" containerName="placement-db-sync" Mar 17 16:03:23 crc kubenswrapper[4767]: E0317 16:03:23.597102 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2af7040-49bc-4689-8af6-a7e3a37db7f2" containerName="keystone-bootstrap" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.597111 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2af7040-49bc-4689-8af6-a7e3a37db7f2" containerName="keystone-bootstrap" Mar 17 16:03:23 crc kubenswrapper[4767]: E0317 16:03:23.597131 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fd01229-44e5-40e9-ba3c-d672309ca045" containerName="dnsmasq-dns" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.597140 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fd01229-44e5-40e9-ba3c-d672309ca045" containerName="dnsmasq-dns" Mar 17 16:03:23 crc kubenswrapper[4767]: E0317 16:03:23.597161 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fd01229-44e5-40e9-ba3c-d672309ca045" containerName="init" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.597187 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fd01229-44e5-40e9-ba3c-d672309ca045" containerName="init" Mar 17 16:03:23 crc kubenswrapper[4767]: E0317 16:03:23.597231 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850bce7d-74fa-4ead-b6c2-fcf21c3083ec" containerName="barbican-db-sync" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.597243 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="850bce7d-74fa-4ead-b6c2-fcf21c3083ec" containerName="barbican-db-sync" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.597699 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5b58621-27a1-479d-a90c-aa96400e1116" containerName="placement-db-sync" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.597745 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="850bce7d-74fa-4ead-b6c2-fcf21c3083ec" containerName="barbican-db-sync" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.597758 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2af7040-49bc-4689-8af6-a7e3a37db7f2" containerName="keystone-bootstrap" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.597775 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fd01229-44e5-40e9-ba3c-d672309ca045" containerName="dnsmasq-dns" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.599065 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.605590 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9kzs7" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.605909 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.622924 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-79f98b8574-mc4pp"] Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.642633 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.642931 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.643089 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.643265 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.647557 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-scripts\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.647669 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pghpv\" (UniqueName: \"kubernetes.io/projected/30d9da8d-1d0e-406a-b87e-aef02d1893cc-kube-api-access-pghpv\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.647714 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-public-tls-certs\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.647758 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-credential-keys\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.647896 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-fernet-keys\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.647971 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-internal-tls-certs\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.648029 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-config-data\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.648119 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-combined-ca-bundle\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.743367 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17172fd-996e-496e-92b5-9cd19edc65f4","Type":"ContainerStarted","Data":"8bf7818523bacace760493e89ff454ca91e19b8de2463beb219f503b023448e1"} Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.752001 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-combined-ca-bundle\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.752147 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.752210 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-h7hgh" event={"ID":"4fd01229-44e5-40e9-ba3c-d672309ca045","Type":"ContainerDied","Data":"5b09d51e9aae6c4d80a24cb7ce2265e67243b148da3da69429bc3f1468141ac4"} Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.752256 4767 scope.go:117] "RemoveContainer" containerID="a58959ac80935fea8311b60a5217d0be5b31501cd9709bf2e61a38d6101605aa" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.752565 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-scripts\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.752687 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pghpv\" (UniqueName: \"kubernetes.io/projected/30d9da8d-1d0e-406a-b87e-aef02d1893cc-kube-api-access-pghpv\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.752746 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-public-tls-certs\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.752811 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-credential-keys\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.752872 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-fernet-keys\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.752906 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-internal-tls-certs\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.752969 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-config-data\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.774225 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-combined-ca-bundle\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.774329 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-config-data\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.774961 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-credential-keys\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.778891 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-fernet-keys\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.779516 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-internal-tls-certs\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.781776 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-public-tls-certs\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.792673 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30d9da8d-1d0e-406a-b87e-aef02d1893cc-scripts\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.833122 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pghpv\" (UniqueName: \"kubernetes.io/projected/30d9da8d-1d0e-406a-b87e-aef02d1893cc-kube-api-access-pghpv\") pod \"keystone-79f98b8574-mc4pp\" (UID: \"30d9da8d-1d0e-406a-b87e-aef02d1893cc\") " pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.896079 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-844b76bcb8-bt49d"] Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.898370 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.912272 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.912703 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bjvk2" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.913628 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.913871 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.914061 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 17 16:03:23 crc kubenswrapper[4767]: I0317 16:03:23.934039 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-844b76bcb8-bt49d"] Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:23.972984 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-public-tls-certs\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.202073 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-combined-ca-bundle\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.202158 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-config-data\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.202350 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-scripts\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.202395 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qflp\" (UniqueName: \"kubernetes.io/projected/7d518ac2-8f5c-4520-b3b4-8e01f6678974-kube-api-access-6qflp\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.202423 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-internal-tls-certs\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.202463 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d518ac2-8f5c-4520-b3b4-8e01f6678974-logs\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.231299 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.330091 4767 scope.go:117] "RemoveContainer" containerID="5cfd31f76b2969652897d667470f38b4cfd75254f1d21c04073a4cb121e9fd73" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.340870 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-internal-tls-certs\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.340935 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d518ac2-8f5c-4520-b3b4-8e01f6678974-logs\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.341019 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-public-tls-certs\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.341196 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-combined-ca-bundle\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.341225 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-config-data\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.341311 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-scripts\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.341348 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qflp\" (UniqueName: \"kubernetes.io/projected/7d518ac2-8f5c-4520-b3b4-8e01f6678974-kube-api-access-6qflp\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.346668 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-config-data\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.347836 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-public-tls-certs\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.366582 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-scripts\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.367579 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d518ac2-8f5c-4520-b3b4-8e01f6678974-logs\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.395446 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-combined-ca-bundle\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.403684 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-h7hgh"] Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.404976 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qflp\" (UniqueName: \"kubernetes.io/projected/7d518ac2-8f5c-4520-b3b4-8e01f6678974-kube-api-access-6qflp\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.419995 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-internal-tls-certs\") pod \"placement-844b76bcb8-bt49d\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.479737 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7f6f54698c-rrhd8"] Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.482272 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.531890 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-h7hgh"] Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.535344 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.542964 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7f6f54698c-rrhd8"] Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.550330 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc95699d-4853-4d4b-9239-019601fd7df8-logs\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.550386 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-combined-ca-bundle\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.550414 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.550513 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn29m\" (UniqueName: \"kubernetes.io/projected/dc95699d-4853-4d4b-9239-019601fd7df8-kube-api-access-xn29m\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.550599 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data-custom\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.561607 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.562215 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.562413 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rnsnz" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.562911 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-579bdf4fdb-kq7rd"] Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.590322 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-579bdf4fdb-kq7rd"] Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.590469 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.610113 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.653160 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data-custom\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.653231 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-logs\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.653299 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc95699d-4853-4d4b-9239-019601fd7df8-logs\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.653337 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-combined-ca-bundle\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.653365 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.653395 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-combined-ca-bundle\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.653423 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgr7z\" (UniqueName: \"kubernetes.io/projected/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-kube-api-access-tgr7z\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.653485 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn29m\" (UniqueName: \"kubernetes.io/projected/dc95699d-4853-4d4b-9239-019601fd7df8-kube-api-access-xn29m\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.653544 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data-custom\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.653563 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.654273 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc95699d-4853-4d4b-9239-019601fd7df8-logs\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.675119 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-combined-ca-bundle\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.677667 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data-custom\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.685657 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.722586 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sgpl5"] Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.725163 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.795268 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sgpl5"] Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.796149 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn29m\" (UniqueName: \"kubernetes.io/projected/dc95699d-4853-4d4b-9239-019601fd7df8-kube-api-access-xn29m\") pod \"barbican-worker-7f6f54698c-rrhd8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.800610 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-config\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.800676 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-combined-ca-bundle\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.800734 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgr7z\" (UniqueName: \"kubernetes.io/projected/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-kube-api-access-tgr7z\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.800965 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.801010 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.801081 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-svc\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.801156 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5v7x\" (UniqueName: \"kubernetes.io/projected/718f5a48-c119-4bae-9c9c-72ed544758ee-kube-api-access-q5v7x\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.801283 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.801322 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data-custom\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.801382 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-logs\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.801465 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.822364 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-logs\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.850549 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-combined-ca-bundle\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.851685 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.899863 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgr7z\" (UniqueName: \"kubernetes.io/projected/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-kube-api-access-tgr7z\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.900752 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8557564488-zkt7w"] Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.929392 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data-custom\") pod \"barbican-keystone-listener-579bdf4fdb-kq7rd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.935090 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.954324 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.959759 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8557564488-zkt7w"] Mar 17 16:03:24 crc kubenswrapper[4767]: I0317 16:03:24.990992 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.020808 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.021083 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-svc\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.023875 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.029576 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5v7x\" (UniqueName: \"kubernetes.io/projected/718f5a48-c119-4bae-9c9c-72ed544758ee-kube-api-access-q5v7x\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.029797 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.029999 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.030285 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-config\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.031345 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-config\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.032347 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.039207 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.051546 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-svc\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.067638 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5bb9c859cd-m865n"] Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.078921 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5v7x\" (UniqueName: \"kubernetes.io/projected/718f5a48-c119-4bae-9c9c-72ed544758ee-kube-api-access-q5v7x\") pod \"dnsmasq-dns-85ff748b95-sgpl5\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.091417 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.142980 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-config-data\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.143073 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-combined-ca-bundle\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.147243 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ca1819-d440-40ed-acd5-ee23e7753496-logs\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.147693 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzb29\" (UniqueName: \"kubernetes.io/projected/26ca1819-d440-40ed-acd5-ee23e7753496-kube-api-access-gzb29\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.147809 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-scripts\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.147944 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-public-tls-certs\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.148050 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-internal-tls-certs\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:25 crc kubenswrapper[4767]: I0317 16:03:25.229072 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5bb9c859cd-m865n"] Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:25.810287 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:25.296235 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-scripts\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.222152 4767 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.222865 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="27c143b3-0f5b-4265-a2e2-d4302a2f3c70" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:25.261230 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-scripts\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.269321 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-public-tls-certs\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.269632 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-internal-tls-certs\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.269861 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-config-data\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.269910 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b1eadc-156d-408a-b815-f06b8133595c-combined-ca-bundle\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.269968 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-combined-ca-bundle\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.270059 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b1eadc-156d-408a-b815-f06b8133595c-config-data\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.270315 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0b1eadc-156d-408a-b815-f06b8133595c-config-data-custom\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.270361 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ca1819-d440-40ed-acd5-ee23e7753496-logs\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.270408 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drtzz\" (UniqueName: \"kubernetes.io/projected/a0b1eadc-156d-408a-b815-f06b8133595c-kube-api-access-drtzz\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.283634 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0b1eadc-156d-408a-b815-f06b8133595c-logs\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.289882 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ca1819-d440-40ed-acd5-ee23e7753496-logs\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.295062 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-public-tls-certs\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.301365 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-config-data\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.301641 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzb29\" (UniqueName: \"kubernetes.io/projected/26ca1819-d440-40ed-acd5-ee23e7753496-kube-api-access-gzb29\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.320439 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fd01229-44e5-40e9-ba3c-d672309ca045" path="/var/lib/kubelet/pods/4fd01229-44e5-40e9-ba3c-d672309ca045/volumes" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.321595 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-77c96db94c-ccm45"] Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.323420 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-combined-ca-bundle\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.338827 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77c96db94c-ccm45"] Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.338878 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7874c48754-xp7vx"] Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.341652 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.349538 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26ca1819-d440-40ed-acd5-ee23e7753496-internal-tls-certs\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.358071 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7874c48754-xp7vx"] Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.358235 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.360427 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.380825 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzb29\" (UniqueName: \"kubernetes.io/projected/26ca1819-d440-40ed-acd5-ee23e7753496-kube-api-access-gzb29\") pod \"placement-8557564488-zkt7w\" (UID: \"26ca1819-d440-40ed-acd5-ee23e7753496\") " pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.407330 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b1eadc-156d-408a-b815-f06b8133595c-combined-ca-bundle\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.409033 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e4e010-5d88-49c6-af61-6dcda69dbe97-logs\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.409413 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b1eadc-156d-408a-b815-f06b8133595c-config-data\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.409704 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e4e010-5d88-49c6-af61-6dcda69dbe97-config-data\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.409950 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5e4e010-5d88-49c6-af61-6dcda69dbe97-config-data-custom\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.410372 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj7mb\" (UniqueName: \"kubernetes.io/projected/f5e4e010-5d88-49c6-af61-6dcda69dbe97-kube-api-access-gj7mb\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.410629 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0b1eadc-156d-408a-b815-f06b8133595c-config-data-custom\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.410919 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drtzz\" (UniqueName: \"kubernetes.io/projected/a0b1eadc-156d-408a-b815-f06b8133595c-kube-api-access-drtzz\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.411255 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0b1eadc-156d-408a-b815-f06b8133595c-logs\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.411653 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e4e010-5d88-49c6-af61-6dcda69dbe97-combined-ca-bundle\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.413078 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0b1eadc-156d-408a-b815-f06b8133595c-logs\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.434998 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0b1eadc-156d-408a-b815-f06b8133595c-config-data-custom\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.436670 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b1eadc-156d-408a-b815-f06b8133595c-combined-ca-bundle\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.460758 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drtzz\" (UniqueName: \"kubernetes.io/projected/a0b1eadc-156d-408a-b815-f06b8133595c-kube-api-access-drtzz\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.463506 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b1eadc-156d-408a-b815-f06b8133595c-config-data\") pod \"barbican-keystone-listener-5bb9c859cd-m865n\" (UID: \"a0b1eadc-156d-408a-b815-f06b8133595c\") " pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.516753 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.516832 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qncbc\" (UniqueName: \"kubernetes.io/projected/22f8c42f-134c-47ed-baa8-0f176a362d38-kube-api-access-qncbc\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.516878 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e4e010-5d88-49c6-af61-6dcda69dbe97-logs\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.516929 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e4e010-5d88-49c6-af61-6dcda69dbe97-config-data\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.516963 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5e4e010-5d88-49c6-af61-6dcda69dbe97-config-data-custom\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.517010 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj7mb\" (UniqueName: \"kubernetes.io/projected/f5e4e010-5d88-49c6-af61-6dcda69dbe97-kube-api-access-gj7mb\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.517047 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-combined-ca-bundle\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.517126 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e4e010-5d88-49c6-af61-6dcda69dbe97-combined-ca-bundle\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.517181 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22f8c42f-134c-47ed-baa8-0f176a362d38-logs\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.517208 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data-custom\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.517865 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e4e010-5d88-49c6-af61-6dcda69dbe97-logs\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.540385 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e4e010-5d88-49c6-af61-6dcda69dbe97-config-data\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.546418 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e4e010-5d88-49c6-af61-6dcda69dbe97-combined-ca-bundle\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.556269 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5e4e010-5d88-49c6-af61-6dcda69dbe97-config-data-custom\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.574206 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.585620 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj7mb\" (UniqueName: \"kubernetes.io/projected/f5e4e010-5d88-49c6-af61-6dcda69dbe97-kube-api-access-gj7mb\") pod \"barbican-worker-77c96db94c-ccm45\" (UID: \"f5e4e010-5d88-49c6-af61-6dcda69dbe97\") " pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.621483 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-combined-ca-bundle\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.621634 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22f8c42f-134c-47ed-baa8-0f176a362d38-logs\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.621665 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data-custom\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.621716 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.621747 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qncbc\" (UniqueName: \"kubernetes.io/projected/22f8c42f-134c-47ed-baa8-0f176a362d38-kube-api-access-qncbc\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.626784 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22f8c42f-134c-47ed-baa8-0f176a362d38-logs\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.650593 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.658566 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-combined-ca-bundle\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.660077 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qncbc\" (UniqueName: \"kubernetes.io/projected/22f8c42f-134c-47ed-baa8-0f176a362d38-kube-api-access-qncbc\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.663813 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data-custom\") pod \"barbican-api-7874c48754-xp7vx\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.734871 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.852965 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-657sp" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.873384 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77c96db94c-ccm45" Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.955252 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-combined-ca-bundle\") pod \"5e6a688d-7438-45a1-9aa7-b7042485163f\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.955312 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8kxb\" (UniqueName: \"kubernetes.io/projected/5e6a688d-7438-45a1-9aa7-b7042485163f-kube-api-access-z8kxb\") pod \"5e6a688d-7438-45a1-9aa7-b7042485163f\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.955553 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-config-data\") pod \"5e6a688d-7438-45a1-9aa7-b7042485163f\" (UID: \"5e6a688d-7438-45a1-9aa7-b7042485163f\") " Mar 17 16:03:26 crc kubenswrapper[4767]: I0317 16:03:26.994081 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.025459 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e6a688d-7438-45a1-9aa7-b7042485163f" (UID: "5e6a688d-7438-45a1-9aa7-b7042485163f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.032031 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e6a688d-7438-45a1-9aa7-b7042485163f-kube-api-access-z8kxb" (OuterVolumeSpecName: "kube-api-access-z8kxb") pod "5e6a688d-7438-45a1-9aa7-b7042485163f" (UID: "5e6a688d-7438-45a1-9aa7-b7042485163f"). InnerVolumeSpecName "kube-api-access-z8kxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.062253 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.062294 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8kxb\" (UniqueName: \"kubernetes.io/projected/5e6a688d-7438-45a1-9aa7-b7042485163f-kube-api-access-z8kxb\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.303037 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-config-data" (OuterVolumeSpecName: "config-data") pod "5e6a688d-7438-45a1-9aa7-b7042485163f" (UID: "5e6a688d-7438-45a1-9aa7-b7042485163f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.366966 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e6a688d-7438-45a1-9aa7-b7042485163f-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.379751 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-657sp" Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.403572 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-657sp" event={"ID":"5e6a688d-7438-45a1-9aa7-b7042485163f","Type":"ContainerDied","Data":"8ee5d42cb494ad0292cece8e63bd267fae73fd7589f072159cae76fd6785fe0d"} Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.403646 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ee5d42cb494ad0292cece8e63bd267fae73fd7589f072159cae76fd6785fe0d" Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.424326 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-844b76bcb8-bt49d"] Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.461060 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-79f98b8574-mc4pp"] Mar 17 16:03:27 crc kubenswrapper[4767]: W0317 16:03:27.484350 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d518ac2_8f5c_4520_b3b4_8e01f6678974.slice/crio-6ce5194d8010af588fcab7c32752c462ca0e06ce8137a51c00273edaadf3c9bf WatchSource:0}: Error finding container 6ce5194d8010af588fcab7c32752c462ca0e06ce8137a51c00273edaadf3c9bf: Status 404 returned error can't find the container with id 6ce5194d8010af588fcab7c32752c462ca0e06ce8137a51c00273edaadf3c9bf Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.554841 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-579bdf4fdb-kq7rd"] Mar 17 16:03:27 crc kubenswrapper[4767]: W0317 16:03:27.588444 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff0f3ad0_1b55_4abb_9eb6_c07d82251bdd.slice/crio-6a7698f4045ba558680b51081489fc574b78c028bcaebe433f4425c74b3e6b03 WatchSource:0}: Error finding container 6a7698f4045ba558680b51081489fc574b78c028bcaebe433f4425c74b3e6b03: Status 404 returned error can't find the container with id 6a7698f4045ba558680b51081489fc574b78c028bcaebe433f4425c74b3e6b03 Mar 17 16:03:27 crc kubenswrapper[4767]: I0317 16:03:27.802884 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7f6f54698c-rrhd8"] Mar 17 16:03:28 crc kubenswrapper[4767]: I0317 16:03:28.551962 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" event={"ID":"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd","Type":"ContainerStarted","Data":"6a7698f4045ba558680b51081489fc574b78c028bcaebe433f4425c74b3e6b03"} Mar 17 16:03:28 crc kubenswrapper[4767]: I0317 16:03:28.573972 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f6f54698c-rrhd8" event={"ID":"dc95699d-4853-4d4b-9239-019601fd7df8","Type":"ContainerStarted","Data":"ef67e9fd96b6864bdce50fffd18441a5cc8a6f39f638ebcc13ab497c753751cb"} Mar 17 16:03:28 crc kubenswrapper[4767]: I0317 16:03:28.635813 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-844b76bcb8-bt49d" event={"ID":"7d518ac2-8f5c-4520-b3b4-8e01f6678974","Type":"ContainerStarted","Data":"6ce5194d8010af588fcab7c32752c462ca0e06ce8137a51c00273edaadf3c9bf"} Mar 17 16:03:28 crc kubenswrapper[4767]: I0317 16:03:28.662697 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79f98b8574-mc4pp" event={"ID":"30d9da8d-1d0e-406a-b87e-aef02d1893cc","Type":"ContainerStarted","Data":"09dc834288f176a3ff107e185dee8bab8242600e28a4118f5f0898fd5c42d1af"} Mar 17 16:03:28 crc kubenswrapper[4767]: I0317 16:03:28.746999 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8557564488-zkt7w"] Mar 17 16:03:28 crc kubenswrapper[4767]: I0317 16:03:28.822782 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sgpl5"] Mar 17 16:03:29 crc kubenswrapper[4767]: W0317 16:03:29.087358 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22f8c42f_134c_47ed_baa8_0f176a362d38.slice/crio-0d5fb30ba7c1b39f8564c92407d43f25f27816d4e5a5aa0510e34b97bad77618 WatchSource:0}: Error finding container 0d5fb30ba7c1b39f8564c92407d43f25f27816d4e5a5aa0510e34b97bad77618: Status 404 returned error can't find the container with id 0d5fb30ba7c1b39f8564c92407d43f25f27816d4e5a5aa0510e34b97bad77618 Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.089698 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7874c48754-xp7vx"] Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.141615 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5bb9c859cd-m865n"] Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.195593 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77c96db94c-ccm45"] Mar 17 16:03:29 crc kubenswrapper[4767]: W0317 16:03:29.205523 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5e4e010_5d88_49c6_af61_6dcda69dbe97.slice/crio-d9b4e8e574887ebfe226bc562d373fbfe3795387107031be281aeaadafc09a65 WatchSource:0}: Error finding container d9b4e8e574887ebfe226bc562d373fbfe3795387107031be281aeaadafc09a65: Status 404 returned error can't find the container with id d9b4e8e574887ebfe226bc562d373fbfe3795387107031be281aeaadafc09a65 Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.715016 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-844b76bcb8-bt49d" event={"ID":"7d518ac2-8f5c-4520-b3b4-8e01f6678974","Type":"ContainerStarted","Data":"6f3a2f449fb65e19cac86104789c3ad946517e7dd0bf835c4455fa34962b7df9"} Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.718548 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79f98b8574-mc4pp" event={"ID":"30d9da8d-1d0e-406a-b87e-aef02d1893cc","Type":"ContainerStarted","Data":"2f6dc2ed067d2c8193780b0fd5a3287289bd10913987127a3f7dc55e102728da"} Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.720047 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.723632 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7874c48754-xp7vx" event={"ID":"22f8c42f-134c-47ed-baa8-0f176a362d38","Type":"ContainerStarted","Data":"0d5fb30ba7c1b39f8564c92407d43f25f27816d4e5a5aa0510e34b97bad77618"} Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.725108 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8557564488-zkt7w" event={"ID":"26ca1819-d440-40ed-acd5-ee23e7753496","Type":"ContainerStarted","Data":"f3f5af758910044177d332c4f8f40c9d9790b55a269032f90aef9208febee6d7"} Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.730352 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" event={"ID":"a0b1eadc-156d-408a-b815-f06b8133595c","Type":"ContainerStarted","Data":"5b1e5637f6b9c4f387578091cd84d09239cb8bab69f96bc60ddc9f6e3fa54c7d"} Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.734702 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77c96db94c-ccm45" event={"ID":"f5e4e010-5d88-49c6-af61-6dcda69dbe97","Type":"ContainerStarted","Data":"d9b4e8e574887ebfe226bc562d373fbfe3795387107031be281aeaadafc09a65"} Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.737138 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pdxns" event={"ID":"712803ce-6e47-40a6-bf5c-84e8aea748e4","Type":"ContainerStarted","Data":"9e8adf44c914c054494ac45c8245d85c8be3816851cea224cae9508ce2b26587"} Mar 17 16:03:29 crc kubenswrapper[4767]: I0317 16:03:29.746737 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" event={"ID":"718f5a48-c119-4bae-9c9c-72ed544758ee","Type":"ContainerStarted","Data":"54dec61b8f1f3cbb7d95d23d5348436aae4cab3a396a7a3433084e138d2d84f2"} Mar 17 16:03:30 crc kubenswrapper[4767]: I0317 16:03:30.164447 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-pdxns" podStartSLOduration=9.64202508 podStartE2EDuration="1m10.164427053s" podCreationTimestamp="2026-03-17 16:02:20 +0000 UTC" firstStartedPulling="2026-03-17 16:02:23.939607433 +0000 UTC m=+1535.352923480" lastFinishedPulling="2026-03-17 16:03:24.462009406 +0000 UTC m=+1595.875325453" observedRunningTime="2026-03-17 16:03:29.923857912 +0000 UTC m=+1601.337173969" watchObservedRunningTime="2026-03-17 16:03:30.164427053 +0000 UTC m=+1601.577743100" Mar 17 16:03:30 crc kubenswrapper[4767]: I0317 16:03:30.222565 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-79f98b8574-mc4pp" podStartSLOduration=7.222535255 podStartE2EDuration="7.222535255s" podCreationTimestamp="2026-03-17 16:03:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:29.959050649 +0000 UTC m=+1601.372366706" watchObservedRunningTime="2026-03-17 16:03:30.222535255 +0000 UTC m=+1601.635851302" Mar 17 16:03:30 crc kubenswrapper[4767]: I0317 16:03:30.798941 4767 generic.go:334] "Generic (PLEG): container finished" podID="718f5a48-c119-4bae-9c9c-72ed544758ee" containerID="962e2a724c2df9cf43853d3b90406002a9710291d367f01a12a678eb7ca5b7a3" exitCode=0 Mar 17 16:03:30 crc kubenswrapper[4767]: I0317 16:03:30.799397 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" event={"ID":"718f5a48-c119-4bae-9c9c-72ed544758ee","Type":"ContainerDied","Data":"962e2a724c2df9cf43853d3b90406002a9710291d367f01a12a678eb7ca5b7a3"} Mar 17 16:03:30 crc kubenswrapper[4767]: I0317 16:03:30.805393 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-844b76bcb8-bt49d" event={"ID":"7d518ac2-8f5c-4520-b3b4-8e01f6678974","Type":"ContainerStarted","Data":"1f135306da0aaa11ef00c9f9f2c78a4ba01695e533b490b69523813a529507ce"} Mar 17 16:03:30 crc kubenswrapper[4767]: I0317 16:03:30.805573 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:30 crc kubenswrapper[4767]: I0317 16:03:30.805675 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:30 crc kubenswrapper[4767]: I0317 16:03:30.808481 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8557564488-zkt7w" event={"ID":"26ca1819-d440-40ed-acd5-ee23e7753496","Type":"ContainerStarted","Data":"061ad96f56f1c31c55a78ee44afa93c7451ca0875d1f8f1d5b28403337968684"} Mar 17 16:03:30 crc kubenswrapper[4767]: I0317 16:03:30.812879 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7874c48754-xp7vx" event={"ID":"22f8c42f-134c-47ed-baa8-0f176a362d38","Type":"ContainerStarted","Data":"d66bf9f482f6948bd0dcab70f3bcd2ceecc90f9ef84e29dc11705235ee9f82d5"} Mar 17 16:03:30 crc kubenswrapper[4767]: I0317 16:03:30.929379 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-844b76bcb8-bt49d" podStartSLOduration=7.929352941 podStartE2EDuration="7.929352941s" podCreationTimestamp="2026-03-17 16:03:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:30.872088923 +0000 UTC m=+1602.285404970" watchObservedRunningTime="2026-03-17 16:03:30.929352941 +0000 UTC m=+1602.342668978" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.592309 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6758b74fd-kt69n"] Mar 17 16:03:31 crc kubenswrapper[4767]: E0317 16:03:31.595778 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e6a688d-7438-45a1-9aa7-b7042485163f" containerName="heat-db-sync" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.595822 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e6a688d-7438-45a1-9aa7-b7042485163f" containerName="heat-db-sync" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.596152 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e6a688d-7438-45a1-9aa7-b7042485163f" containerName="heat-db-sync" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.598072 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.603584 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.604650 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.629329 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6758b74fd-kt69n"] Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.640952 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-public-tls-certs\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.641033 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-internal-tls-certs\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.641503 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-combined-ca-bundle\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.641543 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f4d565d-84a7-42a3-819b-ecec1868a344-logs\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.641568 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-config-data-custom\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.642160 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-config-data\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.642263 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m4bk\" (UniqueName: \"kubernetes.io/projected/6f4d565d-84a7-42a3-819b-ecec1868a344-kube-api-access-5m4bk\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.749439 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-config-data\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.749860 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m4bk\" (UniqueName: \"kubernetes.io/projected/6f4d565d-84a7-42a3-819b-ecec1868a344-kube-api-access-5m4bk\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.749928 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-public-tls-certs\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.750012 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-internal-tls-certs\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.750288 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-combined-ca-bundle\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.750327 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f4d565d-84a7-42a3-819b-ecec1868a344-logs\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.750367 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-config-data-custom\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.758469 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f4d565d-84a7-42a3-819b-ecec1868a344-logs\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.766035 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-internal-tls-certs\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.768560 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-config-data-custom\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.770384 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-public-tls-certs\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.770575 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-config-data\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:31 crc kubenswrapper[4767]: I0317 16:03:31.771213 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4d565d-84a7-42a3-819b-ecec1868a344-combined-ca-bundle\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:32 crc kubenswrapper[4767]: I0317 16:03:32.081141 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m4bk\" (UniqueName: \"kubernetes.io/projected/6f4d565d-84a7-42a3-819b-ecec1868a344-kube-api-access-5m4bk\") pod \"barbican-api-6758b74fd-kt69n\" (UID: \"6f4d565d-84a7-42a3-819b-ecec1868a344\") " pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:32 crc kubenswrapper[4767]: I0317 16:03:32.226701 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8557564488-zkt7w" event={"ID":"26ca1819-d440-40ed-acd5-ee23e7753496","Type":"ContainerStarted","Data":"7f6cb3ec373c38cd031e67e399e5d3b0ce3252e5a89f3ee7ec28f1e4abe4134b"} Mar 17 16:03:32 crc kubenswrapper[4767]: I0317 16:03:32.227620 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:32 crc kubenswrapper[4767]: I0317 16:03:32.227770 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:32 crc kubenswrapper[4767]: I0317 16:03:32.241362 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:32 crc kubenswrapper[4767]: I0317 16:03:32.265589 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7874c48754-xp7vx" event={"ID":"22f8c42f-134c-47ed-baa8-0f176a362d38","Type":"ContainerStarted","Data":"fb68641df086d68c697275b20dc2997ea82074be958cf89d4c1031793e6d4fcb"} Mar 17 16:03:32 crc kubenswrapper[4767]: I0317 16:03:32.265973 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:32 crc kubenswrapper[4767]: I0317 16:03:32.266713 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:32 crc kubenswrapper[4767]: I0317 16:03:32.315297 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-8557564488-zkt7w" podStartSLOduration=8.315261091 podStartE2EDuration="8.315261091s" podCreationTimestamp="2026-03-17 16:03:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:32.265200139 +0000 UTC m=+1603.678516196" watchObservedRunningTime="2026-03-17 16:03:32.315261091 +0000 UTC m=+1603.728577158" Mar 17 16:03:32 crc kubenswrapper[4767]: I0317 16:03:32.343823 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7874c48754-xp7vx" podStartSLOduration=7.343798987 podStartE2EDuration="7.343798987s" podCreationTimestamp="2026-03-17 16:03:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:32.316120395 +0000 UTC m=+1603.729436442" watchObservedRunningTime="2026-03-17 16:03:32.343798987 +0000 UTC m=+1603.757115034" Mar 17 16:03:33 crc kubenswrapper[4767]: I0317 16:03:33.440411 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.076479 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-67c55f74cc-tpfc4"] Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.077094 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-67c55f74cc-tpfc4" podUID="16264c79-b41d-43e1-a692-e084ae52e928" containerName="neutron-api" containerID="cri-o://0806cabe6f745515e02e3ee6523619eef04a5837a0805c88be28d9b6873fca60" gracePeriod=30 Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.077905 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-67c55f74cc-tpfc4" podUID="16264c79-b41d-43e1-a692-e084ae52e928" containerName="neutron-httpd" containerID="cri-o://359eb33421735c4d333f5eede05ad9d61866eb0e09e62023bb64790d496a4e82" gracePeriod=30 Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.094770 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-64d5cdf6bf-rvspf"] Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.101689 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.119110 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64d5cdf6bf-rvspf"] Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.166047 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.166115 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.166432 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghmlc\" (UniqueName: \"kubernetes.io/projected/6e1628c8-705f-4ff0-87e0-231135165322-kube-api-access-ghmlc\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.166515 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-public-tls-certs\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.166562 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-httpd-config\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.166608 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-combined-ca-bundle\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.166650 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-config\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.166704 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-internal-tls-certs\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.166773 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-ovndb-tls-certs\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.268929 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-config\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.269056 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-internal-tls-certs\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.269158 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-ovndb-tls-certs\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.269316 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghmlc\" (UniqueName: \"kubernetes.io/projected/6e1628c8-705f-4ff0-87e0-231135165322-kube-api-access-ghmlc\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.269348 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-public-tls-certs\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.269428 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-httpd-config\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.269640 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-combined-ca-bundle\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.277594 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-httpd-config\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.277996 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-config\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.300345 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-combined-ca-bundle\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.301099 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-ovndb-tls-certs\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.303190 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-internal-tls-certs\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.304956 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e1628c8-705f-4ff0-87e0-231135165322-public-tls-certs\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.305939 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghmlc\" (UniqueName: \"kubernetes.io/projected/6e1628c8-705f-4ff0-87e0-231135165322-kube-api-access-ghmlc\") pod \"neutron-64d5cdf6bf-rvspf\" (UID: \"6e1628c8-705f-4ff0-87e0-231135165322\") " pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.391840 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:34 crc kubenswrapper[4767]: I0317 16:03:34.445063 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:35 crc kubenswrapper[4767]: I0317 16:03:35.516691 4767 generic.go:334] "Generic (PLEG): container finished" podID="16264c79-b41d-43e1-a692-e084ae52e928" containerID="359eb33421735c4d333f5eede05ad9d61866eb0e09e62023bb64790d496a4e82" exitCode=0 Mar 17 16:03:35 crc kubenswrapper[4767]: I0317 16:03:35.517818 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c55f74cc-tpfc4" event={"ID":"16264c79-b41d-43e1-a692-e084ae52e928","Type":"ContainerDied","Data":"359eb33421735c4d333f5eede05ad9d61866eb0e09e62023bb64790d496a4e82"} Mar 17 16:03:35 crc kubenswrapper[4767]: I0317 16:03:35.684475 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6758b74fd-kt69n"] Mar 17 16:03:35 crc kubenswrapper[4767]: I0317 16:03:35.941225 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64d5cdf6bf-rvspf"] Mar 17 16:03:35 crc kubenswrapper[4767]: W0317 16:03:35.969635 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e1628c8_705f_4ff0_87e0_231135165322.slice/crio-503e651b68b97be4ba6b507e9492309168fef936de03a89a2b9ad692a0b4523a WatchSource:0}: Error finding container 503e651b68b97be4ba6b507e9492309168fef936de03a89a2b9ad692a0b4523a: Status 404 returned error can't find the container with id 503e651b68b97be4ba6b507e9492309168fef936de03a89a2b9ad692a0b4523a Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.554917 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" event={"ID":"718f5a48-c119-4bae-9c9c-72ed544758ee","Type":"ContainerStarted","Data":"7cbcbb684b7117829363a6d6637dc12889f3744a6903dce8fbfb35e6f52bd8f8"} Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.558832 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.582754 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64d5cdf6bf-rvspf" event={"ID":"6e1628c8-705f-4ff0-87e0-231135165322","Type":"ContainerStarted","Data":"503e651b68b97be4ba6b507e9492309168fef936de03a89a2b9ad692a0b4523a"} Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.607865 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6758b74fd-kt69n" event={"ID":"6f4d565d-84a7-42a3-819b-ecec1868a344","Type":"ContainerStarted","Data":"9e871759c1059a39ec04eb385b0e30b7103ca01e1ebc158262fc159fba9a843f"} Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.607922 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6758b74fd-kt69n" event={"ID":"6f4d565d-84a7-42a3-819b-ecec1868a344","Type":"ContainerStarted","Data":"f25d9b1cedc76fa5bd9a65a92a2f453acb2031783c3ecc84501d9a1dda6d3012"} Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.644999 4767 generic.go:334] "Generic (PLEG): container finished" podID="16264c79-b41d-43e1-a692-e084ae52e928" containerID="0806cabe6f745515e02e3ee6523619eef04a5837a0805c88be28d9b6873fca60" exitCode=0 Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.645566 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c55f74cc-tpfc4" event={"ID":"16264c79-b41d-43e1-a692-e084ae52e928","Type":"ContainerDied","Data":"0806cabe6f745515e02e3ee6523619eef04a5837a0805c88be28d9b6873fca60"} Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.661431 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" event={"ID":"a0b1eadc-156d-408a-b815-f06b8133595c","Type":"ContainerStarted","Data":"11cc865deb1369000cdf08be26145af45652f0b039b82034da00459c0f7ea39d"} Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.670462 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" event={"ID":"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd","Type":"ContainerStarted","Data":"eafb36bee52ba0fdd7580179e1afc8a82c42db16efb072a002849fddd3ae245d"} Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.673211 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f6f54698c-rrhd8" event={"ID":"dc95699d-4853-4d4b-9239-019601fd7df8","Type":"ContainerStarted","Data":"295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c"} Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.683475 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77c96db94c-ccm45" event={"ID":"f5e4e010-5d88-49c6-af61-6dcda69dbe97","Type":"ContainerStarted","Data":"b1a2420bbb2c79cf3bc1646cf9dcd25677bf2cb6b9acafb2db0e386d3e265dfe"} Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.784749 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.814265 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" podStartSLOduration=12.814231458 podStartE2EDuration="12.814231458s" podCreationTimestamp="2026-03-17 16:03:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:36.611855839 +0000 UTC m=+1608.025171906" watchObservedRunningTime="2026-03-17 16:03:36.814231458 +0000 UTC m=+1608.227547515" Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.921939 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-combined-ca-bundle\") pod \"16264c79-b41d-43e1-a692-e084ae52e928\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.922006 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-config\") pod \"16264c79-b41d-43e1-a692-e084ae52e928\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.922097 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2kzk\" (UniqueName: \"kubernetes.io/projected/16264c79-b41d-43e1-a692-e084ae52e928-kube-api-access-s2kzk\") pod \"16264c79-b41d-43e1-a692-e084ae52e928\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.922205 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-httpd-config\") pod \"16264c79-b41d-43e1-a692-e084ae52e928\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.922252 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-public-tls-certs\") pod \"16264c79-b41d-43e1-a692-e084ae52e928\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.922324 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-internal-tls-certs\") pod \"16264c79-b41d-43e1-a692-e084ae52e928\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.922577 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-ovndb-tls-certs\") pod \"16264c79-b41d-43e1-a692-e084ae52e928\" (UID: \"16264c79-b41d-43e1-a692-e084ae52e928\") " Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.932833 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "16264c79-b41d-43e1-a692-e084ae52e928" (UID: "16264c79-b41d-43e1-a692-e084ae52e928"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:36 crc kubenswrapper[4767]: I0317 16:03:36.933153 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16264c79-b41d-43e1-a692-e084ae52e928-kube-api-access-s2kzk" (OuterVolumeSpecName: "kube-api-access-s2kzk") pod "16264c79-b41d-43e1-a692-e084ae52e928" (UID: "16264c79-b41d-43e1-a692-e084ae52e928"). InnerVolumeSpecName "kube-api-access-s2kzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.034427 4767 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.034822 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2kzk\" (UniqueName: \"kubernetes.io/projected/16264c79-b41d-43e1-a692-e084ae52e928-kube-api-access-s2kzk\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.317043 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "16264c79-b41d-43e1-a692-e084ae52e928" (UID: "16264c79-b41d-43e1-a692-e084ae52e928"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.343532 4767 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.608573 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16264c79-b41d-43e1-a692-e084ae52e928" (UID: "16264c79-b41d-43e1-a692-e084ae52e928"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.621933 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "16264c79-b41d-43e1-a692-e084ae52e928" (UID: "16264c79-b41d-43e1-a692-e084ae52e928"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.621986 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "16264c79-b41d-43e1-a692-e084ae52e928" (UID: "16264c79-b41d-43e1-a692-e084ae52e928"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.625827 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-config" (OuterVolumeSpecName: "config") pod "16264c79-b41d-43e1-a692-e084ae52e928" (UID: "16264c79-b41d-43e1-a692-e084ae52e928"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.724404 4767 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.724653 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.724676 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.724693 4767 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16264c79-b41d-43e1-a692-e084ae52e928-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.726402 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6758b74fd-kt69n" event={"ID":"6f4d565d-84a7-42a3-819b-ecec1868a344","Type":"ContainerStarted","Data":"9571203a8a59719c8a4455ce831a0f150e4c4f9ea439ec4b19cc48198cfcb10b"} Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.726493 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.726539 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.747376 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c55f74cc-tpfc4" event={"ID":"16264c79-b41d-43e1-a692-e084ae52e928","Type":"ContainerDied","Data":"2f1b9b811014cbdaba721aa6146c407c4153b6cf9076e2b9678adc5f0af21d0f"} Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.747532 4767 scope.go:117] "RemoveContainer" containerID="359eb33421735c4d333f5eede05ad9d61866eb0e09e62023bb64790d496a4e82" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.747864 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67c55f74cc-tpfc4" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.773928 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6758b74fd-kt69n" podStartSLOduration=6.773899386 podStartE2EDuration="6.773899386s" podCreationTimestamp="2026-03-17 16:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:37.756275702 +0000 UTC m=+1609.169591759" watchObservedRunningTime="2026-03-17 16:03:37.773899386 +0000 UTC m=+1609.187215433" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.785420 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" event={"ID":"a0b1eadc-156d-408a-b815-f06b8133595c","Type":"ContainerStarted","Data":"52010c76831cafb343c6e8f4bf461dff94b933e79c9a1193d8032c8118753491"} Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.807640 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" event={"ID":"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd","Type":"ContainerStarted","Data":"2055be26e06e56bc2b2a7957901aafdce92a48fee3057028796f99ee80231978"} Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.835427 4767 scope.go:117] "RemoveContainer" containerID="0806cabe6f745515e02e3ee6523619eef04a5837a0805c88be28d9b6873fca60" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.852243 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f6f54698c-rrhd8" event={"ID":"dc95699d-4853-4d4b-9239-019601fd7df8","Type":"ContainerStarted","Data":"1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325"} Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.868305 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-67c55f74cc-tpfc4"] Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.887521 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77c96db94c-ccm45" event={"ID":"f5e4e010-5d88-49c6-af61-6dcda69dbe97","Type":"ContainerStarted","Data":"87c9951c644f0e55cdd4fc8d8a182bba42f15fe7592491306807a407e9332a5d"} Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.902373 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-67c55f74cc-tpfc4"] Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.906108 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64d5cdf6bf-rvspf" event={"ID":"6e1628c8-705f-4ff0-87e0-231135165322","Type":"ContainerStarted","Data":"e39c59d60a86ef9afa57f54bf42ae4825655d4ac49c06278f78d4aea9b64e89d"} Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.911275 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5bb9c859cd-m865n" podStartSLOduration=7.894590107 podStartE2EDuration="13.900787575s" podCreationTimestamp="2026-03-17 16:03:24 +0000 UTC" firstStartedPulling="2026-03-17 16:03:29.197096025 +0000 UTC m=+1600.610412062" lastFinishedPulling="2026-03-17 16:03:35.203293483 +0000 UTC m=+1606.616609530" observedRunningTime="2026-03-17 16:03:37.83664815 +0000 UTC m=+1609.249964197" watchObservedRunningTime="2026-03-17 16:03:37.900787575 +0000 UTC m=+1609.314103642" Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.984232 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-579bdf4fdb-kq7rd"] Mar 17 16:03:37 crc kubenswrapper[4767]: I0317 16:03:37.986060 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" podStartSLOduration=6.397827639 podStartE2EDuration="13.986040953s" podCreationTimestamp="2026-03-17 16:03:24 +0000 UTC" firstStartedPulling="2026-03-17 16:03:27.605721521 +0000 UTC m=+1599.019037568" lastFinishedPulling="2026-03-17 16:03:35.193934835 +0000 UTC m=+1606.607250882" observedRunningTime="2026-03-17 16:03:37.87647247 +0000 UTC m=+1609.289788517" watchObservedRunningTime="2026-03-17 16:03:37.986040953 +0000 UTC m=+1609.399357030" Mar 17 16:03:38 crc kubenswrapper[4767]: I0317 16:03:38.013037 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7f6f54698c-rrhd8" podStartSLOduration=7.825661066 podStartE2EDuration="15.013007385s" podCreationTimestamp="2026-03-17 16:03:23 +0000 UTC" firstStartedPulling="2026-03-17 16:03:28.015193652 +0000 UTC m=+1599.428509699" lastFinishedPulling="2026-03-17 16:03:35.202539971 +0000 UTC m=+1606.615856018" observedRunningTime="2026-03-17 16:03:37.913636583 +0000 UTC m=+1609.326952630" watchObservedRunningTime="2026-03-17 16:03:38.013007385 +0000 UTC m=+1609.426323442" Mar 17 16:03:38 crc kubenswrapper[4767]: I0317 16:03:38.040936 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-77c96db94c-ccm45" podStartSLOduration=8.17224272 podStartE2EDuration="14.040905763s" podCreationTimestamp="2026-03-17 16:03:24 +0000 UTC" firstStartedPulling="2026-03-17 16:03:29.267100298 +0000 UTC m=+1600.680416345" lastFinishedPulling="2026-03-17 16:03:35.135763341 +0000 UTC m=+1606.549079388" observedRunningTime="2026-03-17 16:03:37.951424893 +0000 UTC m=+1609.364740940" watchObservedRunningTime="2026-03-17 16:03:38.040905763 +0000 UTC m=+1609.454221810" Mar 17 16:03:38 crc kubenswrapper[4767]: I0317 16:03:38.072854 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7f6f54698c-rrhd8"] Mar 17 16:03:38 crc kubenswrapper[4767]: I0317 16:03:38.936002 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64d5cdf6bf-rvspf" event={"ID":"6e1628c8-705f-4ff0-87e0-231135165322","Type":"ContainerStarted","Data":"a2306b580c129f806563d865601b617814059ece4abfda29d392817c698aaed3"} Mar 17 16:03:38 crc kubenswrapper[4767]: I0317 16:03:38.938321 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:03:38 crc kubenswrapper[4767]: I0317 16:03:38.973004 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-64d5cdf6bf-rvspf" podStartSLOduration=4.972980952 podStartE2EDuration="4.972980952s" podCreationTimestamp="2026-03-17 16:03:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:38.966664071 +0000 UTC m=+1610.379980118" watchObservedRunningTime="2026-03-17 16:03:38.972980952 +0000 UTC m=+1610.386296999" Mar 17 16:03:39 crc kubenswrapper[4767]: I0317 16:03:39.393114 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16264c79-b41d-43e1-a692-e084ae52e928" path="/var/lib/kubelet/pods/16264c79-b41d-43e1-a692-e084ae52e928/volumes" Mar 17 16:03:39 crc kubenswrapper[4767]: I0317 16:03:39.961236 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" podUID="ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" containerName="barbican-keystone-listener-log" containerID="cri-o://eafb36bee52ba0fdd7580179e1afc8a82c42db16efb072a002849fddd3ae245d" gracePeriod=30 Mar 17 16:03:39 crc kubenswrapper[4767]: I0317 16:03:39.961393 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" podUID="ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" containerName="barbican-keystone-listener" containerID="cri-o://2055be26e06e56bc2b2a7957901aafdce92a48fee3057028796f99ee80231978" gracePeriod=30 Mar 17 16:03:39 crc kubenswrapper[4767]: I0317 16:03:39.961841 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7f6f54698c-rrhd8" podUID="dc95699d-4853-4d4b-9239-019601fd7df8" containerName="barbican-worker-log" containerID="cri-o://295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c" gracePeriod=30 Mar 17 16:03:39 crc kubenswrapper[4767]: I0317 16:03:39.961893 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7f6f54698c-rrhd8" podUID="dc95699d-4853-4d4b-9239-019601fd7df8" containerName="barbican-worker" containerID="cri-o://1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325" gracePeriod=30 Mar 17 16:03:40 crc kubenswrapper[4767]: I0317 16:03:40.679763 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:40 crc kubenswrapper[4767]: I0317 16:03:40.817402 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:03:40 crc kubenswrapper[4767]: I0317 16:03:40.832828 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:41 crc kubenswrapper[4767]: I0317 16:03:41.261766 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7874c48754-xp7vx" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.209:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 16:03:41 crc kubenswrapper[4767]: I0317 16:03:41.362211 4767 generic.go:334] "Generic (PLEG): container finished" podID="ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" containerID="eafb36bee52ba0fdd7580179e1afc8a82c42db16efb072a002849fddd3ae245d" exitCode=143 Mar 17 16:03:41 crc kubenswrapper[4767]: I0317 16:03:41.390872 4767 generic.go:334] "Generic (PLEG): container finished" podID="dc95699d-4853-4d4b-9239-019601fd7df8" containerID="295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c" exitCode=143 Mar 17 16:03:41 crc kubenswrapper[4767]: I0317 16:03:41.412743 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" event={"ID":"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd","Type":"ContainerDied","Data":"eafb36bee52ba0fdd7580179e1afc8a82c42db16efb072a002849fddd3ae245d"} Mar 17 16:03:41 crc kubenswrapper[4767]: I0317 16:03:41.412811 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f6f54698c-rrhd8" event={"ID":"dc95699d-4853-4d4b-9239-019601fd7df8","Type":"ContainerDied","Data":"295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c"} Mar 17 16:03:41 crc kubenswrapper[4767]: I0317 16:03:41.427234 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kh22t"] Mar 17 16:03:41 crc kubenswrapper[4767]: I0317 16:03:41.427565 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" podUID="b7d5a577-7603-4cfe-a0c9-a899eeaab93d" containerName="dnsmasq-dns" containerID="cri-o://0f423a7202aa9a2dde93bb3217ea9dd20ec328c78d33d70274b1ec90128c1472" gracePeriod=10 Mar 17 16:03:41 crc kubenswrapper[4767]: I0317 16:03:41.710629 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" podUID="b7d5a577-7603-4cfe-a0c9-a899eeaab93d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.198:5353: connect: connection refused" Mar 17 16:03:42 crc kubenswrapper[4767]: I0317 16:03:42.497924 4767 generic.go:334] "Generic (PLEG): container finished" podID="712803ce-6e47-40a6-bf5c-84e8aea748e4" containerID="9e8adf44c914c054494ac45c8245d85c8be3816851cea224cae9508ce2b26587" exitCode=0 Mar 17 16:03:42 crc kubenswrapper[4767]: I0317 16:03:42.498244 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pdxns" event={"ID":"712803ce-6e47-40a6-bf5c-84e8aea748e4","Type":"ContainerDied","Data":"9e8adf44c914c054494ac45c8245d85c8be3816851cea224cae9508ce2b26587"} Mar 17 16:03:42 crc kubenswrapper[4767]: I0317 16:03:42.509546 4767 generic.go:334] "Generic (PLEG): container finished" podID="b7d5a577-7603-4cfe-a0c9-a899eeaab93d" containerID="0f423a7202aa9a2dde93bb3217ea9dd20ec328c78d33d70274b1ec90128c1472" exitCode=0 Mar 17 16:03:42 crc kubenswrapper[4767]: I0317 16:03:42.509596 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" event={"ID":"b7d5a577-7603-4cfe-a0c9-a899eeaab93d","Type":"ContainerDied","Data":"0f423a7202aa9a2dde93bb3217ea9dd20ec328c78d33d70274b1ec90128c1472"} Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.173335 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-976fc"] Mar 17 16:03:44 crc kubenswrapper[4767]: E0317 16:03:44.175780 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16264c79-b41d-43e1-a692-e084ae52e928" containerName="neutron-api" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.175845 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="16264c79-b41d-43e1-a692-e084ae52e928" containerName="neutron-api" Mar 17 16:03:44 crc kubenswrapper[4767]: E0317 16:03:44.175913 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16264c79-b41d-43e1-a692-e084ae52e928" containerName="neutron-httpd" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.175924 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="16264c79-b41d-43e1-a692-e084ae52e928" containerName="neutron-httpd" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.176476 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="16264c79-b41d-43e1-a692-e084ae52e928" containerName="neutron-api" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.176498 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="16264c79-b41d-43e1-a692-e084ae52e928" containerName="neutron-httpd" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.179331 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-976fc" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.260959 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5nz6\" (UniqueName: \"kubernetes.io/projected/25debbd1-41e4-42db-af74-c058c0d6d9b2-kube-api-access-m5nz6\") pod \"community-operators-976fc\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " pod="openshift-marketplace/community-operators-976fc" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.261619 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-catalog-content\") pod \"community-operators-976fc\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " pod="openshift-marketplace/community-operators-976fc" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.261705 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-utilities\") pod \"community-operators-976fc\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " pod="openshift-marketplace/community-operators-976fc" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.261955 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-976fc"] Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.372499 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5nz6\" (UniqueName: \"kubernetes.io/projected/25debbd1-41e4-42db-af74-c058c0d6d9b2-kube-api-access-m5nz6\") pod \"community-operators-976fc\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " pod="openshift-marketplace/community-operators-976fc" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.379267 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-catalog-content\") pod \"community-operators-976fc\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " pod="openshift-marketplace/community-operators-976fc" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.382267 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-utilities\") pod \"community-operators-976fc\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " pod="openshift-marketplace/community-operators-976fc" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.382416 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-catalog-content\") pod \"community-operators-976fc\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " pod="openshift-marketplace/community-operators-976fc" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.383181 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-utilities\") pod \"community-operators-976fc\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " pod="openshift-marketplace/community-operators-976fc" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.412930 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5nz6\" (UniqueName: \"kubernetes.io/projected/25debbd1-41e4-42db-af74-c058c0d6d9b2-kube-api-access-m5nz6\") pod \"community-operators-976fc\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " pod="openshift-marketplace/community-operators-976fc" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.528090 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-976fc" Mar 17 16:03:44 crc kubenswrapper[4767]: I0317 16:03:44.885601 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:45 crc kubenswrapper[4767]: I0317 16:03:45.859134 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6758b74fd-kt69n" Mar 17 16:03:45 crc kubenswrapper[4767]: I0317 16:03:45.984571 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7874c48754-xp7vx"] Mar 17 16:03:45 crc kubenswrapper[4767]: I0317 16:03:45.984907 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7874c48754-xp7vx" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerName="barbican-api-log" containerID="cri-o://d66bf9f482f6948bd0dcab70f3bcd2ceecc90f9ef84e29dc11705235ee9f82d5" gracePeriod=30 Mar 17 16:03:45 crc kubenswrapper[4767]: I0317 16:03:45.985480 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7874c48754-xp7vx" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerName="barbican-api" containerID="cri-o://fb68641df086d68c697275b20dc2997ea82074be958cf89d4c1031793e6d4fcb" gracePeriod=30 Mar 17 16:03:46 crc kubenswrapper[4767]: I0317 16:03:46.599789 4767 generic.go:334] "Generic (PLEG): container finished" podID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerID="d66bf9f482f6948bd0dcab70f3bcd2ceecc90f9ef84e29dc11705235ee9f82d5" exitCode=143 Mar 17 16:03:46 crc kubenswrapper[4767]: I0317 16:03:46.599896 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7874c48754-xp7vx" event={"ID":"22f8c42f-134c-47ed-baa8-0f176a362d38","Type":"ContainerDied","Data":"d66bf9f482f6948bd0dcab70f3bcd2ceecc90f9ef84e29dc11705235ee9f82d5"} Mar 17 16:03:46 crc kubenswrapper[4767]: I0317 16:03:46.709758 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" podUID="b7d5a577-7603-4cfe-a0c9-a899eeaab93d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.198:5353: connect: connection refused" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.590770 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pdxns" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.632781 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pdxns" event={"ID":"712803ce-6e47-40a6-bf5c-84e8aea748e4","Type":"ContainerDied","Data":"c22c642712671e1561ca6caa6e4eba8e87c35ff51dce05cbc0c14b6a680aa990"} Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.632846 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c22c642712671e1561ca6caa6e4eba8e87c35ff51dce05cbc0c14b6a680aa990" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.632916 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pdxns" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.761902 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-config-data\") pod \"712803ce-6e47-40a6-bf5c-84e8aea748e4\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.761994 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jmsj\" (UniqueName: \"kubernetes.io/projected/712803ce-6e47-40a6-bf5c-84e8aea748e4-kube-api-access-6jmsj\") pod \"712803ce-6e47-40a6-bf5c-84e8aea748e4\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.762985 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-scripts\") pod \"712803ce-6e47-40a6-bf5c-84e8aea748e4\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.763120 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-combined-ca-bundle\") pod \"712803ce-6e47-40a6-bf5c-84e8aea748e4\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.763139 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/712803ce-6e47-40a6-bf5c-84e8aea748e4-etc-machine-id\") pod \"712803ce-6e47-40a6-bf5c-84e8aea748e4\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.763243 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-db-sync-config-data\") pod \"712803ce-6e47-40a6-bf5c-84e8aea748e4\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.763439 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/712803ce-6e47-40a6-bf5c-84e8aea748e4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "712803ce-6e47-40a6-bf5c-84e8aea748e4" (UID: "712803ce-6e47-40a6-bf5c-84e8aea748e4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.764461 4767 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/712803ce-6e47-40a6-bf5c-84e8aea748e4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.769384 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/712803ce-6e47-40a6-bf5c-84e8aea748e4-kube-api-access-6jmsj" (OuterVolumeSpecName: "kube-api-access-6jmsj") pod "712803ce-6e47-40a6-bf5c-84e8aea748e4" (UID: "712803ce-6e47-40a6-bf5c-84e8aea748e4"). InnerVolumeSpecName "kube-api-access-6jmsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.771110 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-scripts" (OuterVolumeSpecName: "scripts") pod "712803ce-6e47-40a6-bf5c-84e8aea748e4" (UID: "712803ce-6e47-40a6-bf5c-84e8aea748e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.816181 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "712803ce-6e47-40a6-bf5c-84e8aea748e4" (UID: "712803ce-6e47-40a6-bf5c-84e8aea748e4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.830594 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "712803ce-6e47-40a6-bf5c-84e8aea748e4" (UID: "712803ce-6e47-40a6-bf5c-84e8aea748e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.871730 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.872290 4767 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.872372 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jmsj\" (UniqueName: \"kubernetes.io/projected/712803ce-6e47-40a6-bf5c-84e8aea748e4-kube-api-access-6jmsj\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.872451 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.974153 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-config-data" (OuterVolumeSpecName: "config-data") pod "712803ce-6e47-40a6-bf5c-84e8aea748e4" (UID: "712803ce-6e47-40a6-bf5c-84e8aea748e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.975963 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-config-data\") pod \"712803ce-6e47-40a6-bf5c-84e8aea748e4\" (UID: \"712803ce-6e47-40a6-bf5c-84e8aea748e4\") " Mar 17 16:03:47 crc kubenswrapper[4767]: W0317 16:03:47.976413 4767 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/712803ce-6e47-40a6-bf5c-84e8aea748e4/volumes/kubernetes.io~secret/config-data Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.976446 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-config-data" (OuterVolumeSpecName: "config-data") pod "712803ce-6e47-40a6-bf5c-84e8aea748e4" (UID: "712803ce-6e47-40a6-bf5c-84e8aea748e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:47 crc kubenswrapper[4767]: I0317 16:03:47.977529 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/712803ce-6e47-40a6-bf5c-84e8aea748e4-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:48 crc kubenswrapper[4767]: I0317 16:03:48.888126 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:48 crc kubenswrapper[4767]: I0317 16:03:48.905159 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-config\") pod \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " Mar 17 16:03:48 crc kubenswrapper[4767]: I0317 16:03:48.905265 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-swift-storage-0\") pod \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.009535 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-svc\") pod \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.009703 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-nb\") pod \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.009752 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-sb\") pod \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.009825 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84kkr\" (UniqueName: \"kubernetes.io/projected/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-kube-api-access-84kkr\") pod \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\" (UID: \"b7d5a577-7603-4cfe-a0c9-a899eeaab93d\") " Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.035716 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-config" (OuterVolumeSpecName: "config") pod "b7d5a577-7603-4cfe-a0c9-a899eeaab93d" (UID: "b7d5a577-7603-4cfe-a0c9-a899eeaab93d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.055591 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b7d5a577-7603-4cfe-a0c9-a899eeaab93d" (UID: "b7d5a577-7603-4cfe-a0c9-a899eeaab93d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.068871 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-kube-api-access-84kkr" (OuterVolumeSpecName: "kube-api-access-84kkr") pod "b7d5a577-7603-4cfe-a0c9-a899eeaab93d" (UID: "b7d5a577-7603-4cfe-a0c9-a899eeaab93d"). InnerVolumeSpecName "kube-api-access-84kkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.091060 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 16:03:49 crc kubenswrapper[4767]: E0317 16:03:49.092187 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712803ce-6e47-40a6-bf5c-84e8aea748e4" containerName="cinder-db-sync" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.092292 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="712803ce-6e47-40a6-bf5c-84e8aea748e4" containerName="cinder-db-sync" Mar 17 16:03:49 crc kubenswrapper[4767]: E0317 16:03:49.092368 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7d5a577-7603-4cfe-a0c9-a899eeaab93d" containerName="init" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.092428 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7d5a577-7603-4cfe-a0c9-a899eeaab93d" containerName="init" Mar 17 16:03:49 crc kubenswrapper[4767]: E0317 16:03:49.092506 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7d5a577-7603-4cfe-a0c9-a899eeaab93d" containerName="dnsmasq-dns" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.092587 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7d5a577-7603-4cfe-a0c9-a899eeaab93d" containerName="dnsmasq-dns" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.092924 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7d5a577-7603-4cfe-a0c9-a899eeaab93d" containerName="dnsmasq-dns" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.093032 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="712803ce-6e47-40a6-bf5c-84e8aea748e4" containerName="cinder-db-sync" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.094627 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.099562 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7rfgj" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.099885 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.100005 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.100712 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.113756 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.114140 4767 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.114481 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84kkr\" (UniqueName: \"kubernetes.io/projected/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-kube-api-access-84kkr\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.128358 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.171162 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b7d5a577-7603-4cfe-a0c9-a899eeaab93d" (UID: "b7d5a577-7603-4cfe-a0c9-a899eeaab93d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.210307 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-h4pcz"] Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.510951 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-scripts\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.515790 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb6xw\" (UniqueName: \"kubernetes.io/projected/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-kube-api-access-zb6xw\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.516090 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.516211 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.516295 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.516591 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.516815 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.517112 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.558649 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b7d5a577-7603-4cfe-a0c9-a899eeaab93d" (UID: "b7d5a577-7603-4cfe-a0c9-a899eeaab93d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.590039 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b7d5a577-7603-4cfe-a0c9-a899eeaab93d" (UID: "b7d5a577-7603-4cfe-a0c9-a899eeaab93d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.611532 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-h4pcz"] Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.654689 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.654799 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.654859 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.655040 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.655163 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-scripts\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.655215 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb6xw\" (UniqueName: \"kubernetes.io/projected/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-kube-api-access-zb6xw\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.655375 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.655389 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7d5a577-7603-4cfe-a0c9-a899eeaab93d-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.655420 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.659135 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.659713 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.660066 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.693538 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.705477 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.706632 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.780685 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb6xw\" (UniqueName: \"kubernetes.io/projected/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-kube-api-access-zb6xw\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.797071 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.809026 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.813275 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-scripts\") pod \"cinder-scheduler-0\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " pod="openstack/cinder-scheduler-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.823806 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7874c48754-xp7vx" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.209:9311/healthcheck\": read tcp 10.217.0.2:37508->10.217.0.209:9311: read: connection reset by peer" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.824082 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7874c48754-xp7vx" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.209:9311/healthcheck\": read tcp 10.217.0.2:37498->10.217.0.209:9311: read: connection reset by peer" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.833609 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.839560 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-config\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.839629 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.839741 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f964f8b0-adf7-4bf3-880c-f123640015af-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.839897 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.840030 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data-custom\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.841070 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.841138 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-scripts\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.841194 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np485\" (UniqueName: \"kubernetes.io/projected/2f4160a8-203a-41e6-a90f-f0800ae776b1-kube-api-access-np485\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.841218 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f964f8b0-adf7-4bf3-880c-f123640015af-logs\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.841248 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.841445 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.841525 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgnks\" (UniqueName: \"kubernetes.io/projected/f964f8b0-adf7-4bf3-880c-f123640015af-kube-api-access-dgnks\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.841625 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.841803 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.871777 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" event={"ID":"b7d5a577-7603-4cfe-a0c9-a899eeaab93d","Type":"ContainerDied","Data":"48421ec571d9b3ae98f80a88d66d8c7f416a09279d4899bd338c256b2cd6832e"} Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.871869 4767 scope.go:117] "RemoveContainer" containerID="0f423a7202aa9a2dde93bb3217ea9dd20ec328c78d33d70274b1ec90128c1472" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.872109 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-kh22t" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945129 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945196 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945230 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data-custom\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945257 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945279 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-scripts\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945304 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np485\" (UniqueName: \"kubernetes.io/projected/2f4160a8-203a-41e6-a90f-f0800ae776b1-kube-api-access-np485\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945322 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f964f8b0-adf7-4bf3-880c-f123640015af-logs\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945338 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945359 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945381 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgnks\" (UniqueName: \"kubernetes.io/projected/f964f8b0-adf7-4bf3-880c-f123640015af-kube-api-access-dgnks\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945435 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945534 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-config\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945552 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945585 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f964f8b0-adf7-4bf3-880c-f123640015af-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.945700 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f964f8b0-adf7-4bf3-880c-f123640015af-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.947119 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.947924 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-config\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.948303 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kh22t"] Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.948535 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.948949 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f964f8b0-adf7-4bf3-880c-f123640015af-logs\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.949204 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.951806 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.953887 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data-custom\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.954129 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.961619 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kh22t"] Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.963127 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-scripts\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.976681 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgnks\" (UniqueName: \"kubernetes.io/projected/f964f8b0-adf7-4bf3-880c-f123640015af-kube-api-access-dgnks\") pod \"cinder-api-0\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " pod="openstack/cinder-api-0" Mar 17 16:03:49 crc kubenswrapper[4767]: I0317 16:03:49.983162 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np485\" (UniqueName: \"kubernetes.io/projected/2f4160a8-203a-41e6-a90f-f0800ae776b1-kube-api-access-np485\") pod \"dnsmasq-dns-5c9776ccc5-h4pcz\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:50 crc kubenswrapper[4767]: I0317 16:03:50.026199 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7rfgj" Mar 17 16:03:50 crc kubenswrapper[4767]: I0317 16:03:50.035324 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 16:03:50 crc kubenswrapper[4767]: I0317 16:03:50.287355 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 16:03:50 crc kubenswrapper[4767]: I0317 16:03:50.295965 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:50 crc kubenswrapper[4767]: I0317 16:03:50.901653 4767 generic.go:334] "Generic (PLEG): container finished" podID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerID="fb68641df086d68c697275b20dc2997ea82074be958cf89d4c1031793e6d4fcb" exitCode=0 Mar 17 16:03:50 crc kubenswrapper[4767]: I0317 16:03:50.901877 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7874c48754-xp7vx" event={"ID":"22f8c42f-134c-47ed-baa8-0f176a362d38","Type":"ContainerDied","Data":"fb68641df086d68c697275b20dc2997ea82074be958cf89d4c1031793e6d4fcb"} Mar 17 16:03:50 crc kubenswrapper[4767]: E0317 16:03:50.971985 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Mar 17 16:03:50 crc kubenswrapper[4767]: E0317 16:03:50.972279 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-25njk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(d17172fd-996e-496e-92b5-9cd19edc65f4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 16:03:50 crc kubenswrapper[4767]: E0317 16:03:50.973600 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="d17172fd-996e-496e-92b5-9cd19edc65f4" Mar 17 16:03:50 crc kubenswrapper[4767]: I0317 16:03:50.992409 4767 scope.go:117] "RemoveContainer" containerID="bd16df8e254f555876f90d11edba095368d04cffdb39159ec0dde01b5e308819" Mar 17 16:03:51 crc kubenswrapper[4767]: I0317 16:03:51.392460 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7d5a577-7603-4cfe-a0c9-a899eeaab93d" path="/var/lib/kubelet/pods/b7d5a577-7603-4cfe-a0c9-a899eeaab93d/volumes" Mar 17 16:03:51 crc kubenswrapper[4767]: I0317 16:03:51.939565 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7874c48754-xp7vx" event={"ID":"22f8c42f-134c-47ed-baa8-0f176a362d38","Type":"ContainerDied","Data":"0d5fb30ba7c1b39f8564c92407d43f25f27816d4e5a5aa0510e34b97bad77618"} Mar 17 16:03:51 crc kubenswrapper[4767]: I0317 16:03:51.940057 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d5fb30ba7c1b39f8564c92407d43f25f27816d4e5a5aa0510e34b97bad77618" Mar 17 16:03:51 crc kubenswrapper[4767]: I0317 16:03:51.940314 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d17172fd-996e-496e-92b5-9cd19edc65f4" containerName="ceilometer-notification-agent" containerID="cri-o://94848df684c7d16e35f3425fecf44bad6bb45f6cba749b60673a0a0b7fe71222" gracePeriod=30 Mar 17 16:03:51 crc kubenswrapper[4767]: I0317 16:03:51.940726 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d17172fd-996e-496e-92b5-9cd19edc65f4" containerName="sg-core" containerID="cri-o://8bf7818523bacace760493e89ff454ca91e19b8de2463beb219f503b023448e1" gracePeriod=30 Mar 17 16:03:51 crc kubenswrapper[4767]: I0317 16:03:51.977132 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.190891 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data\") pod \"22f8c42f-134c-47ed-baa8-0f176a362d38\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.191076 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-combined-ca-bundle\") pod \"22f8c42f-134c-47ed-baa8-0f176a362d38\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.191339 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data-custom\") pod \"22f8c42f-134c-47ed-baa8-0f176a362d38\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.191509 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22f8c42f-134c-47ed-baa8-0f176a362d38-logs\") pod \"22f8c42f-134c-47ed-baa8-0f176a362d38\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.194399 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22f8c42f-134c-47ed-baa8-0f176a362d38-logs" (OuterVolumeSpecName: "logs") pod "22f8c42f-134c-47ed-baa8-0f176a362d38" (UID: "22f8c42f-134c-47ed-baa8-0f176a362d38"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.227837 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "22f8c42f-134c-47ed-baa8-0f176a362d38" (UID: "22f8c42f-134c-47ed-baa8-0f176a362d38"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.296973 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qncbc\" (UniqueName: \"kubernetes.io/projected/22f8c42f-134c-47ed-baa8-0f176a362d38-kube-api-access-qncbc\") pod \"22f8c42f-134c-47ed-baa8-0f176a362d38\" (UID: \"22f8c42f-134c-47ed-baa8-0f176a362d38\") " Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.298433 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22f8c42f-134c-47ed-baa8-0f176a362d38-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.298452 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.319619 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22f8c42f-134c-47ed-baa8-0f176a362d38-kube-api-access-qncbc" (OuterVolumeSpecName: "kube-api-access-qncbc") pod "22f8c42f-134c-47ed-baa8-0f176a362d38" (UID: "22f8c42f-134c-47ed-baa8-0f176a362d38"). InnerVolumeSpecName "kube-api-access-qncbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.319854 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22f8c42f-134c-47ed-baa8-0f176a362d38" (UID: "22f8c42f-134c-47ed-baa8-0f176a362d38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.359451 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data" (OuterVolumeSpecName: "config-data") pod "22f8c42f-134c-47ed-baa8-0f176a362d38" (UID: "22f8c42f-134c-47ed-baa8-0f176a362d38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.385918 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-976fc"] Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.404719 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.404758 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22f8c42f-134c-47ed-baa8-0f176a362d38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.404770 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qncbc\" (UniqueName: \"kubernetes.io/projected/22f8c42f-134c-47ed-baa8-0f176a362d38-kube-api-access-qncbc\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.413729 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.578466 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.624205 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 16:03:52 crc kubenswrapper[4767]: I0317 16:03:52.765287 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-h4pcz"] Mar 17 16:03:53 crc kubenswrapper[4767]: I0317 16:03:53.028974 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f964f8b0-adf7-4bf3-880c-f123640015af","Type":"ContainerStarted","Data":"05f44bc81ba5e1bc2ac892fbd1f628a4ca9644b6c12c38a4614e16c3c8c84dec"} Mar 17 16:03:53 crc kubenswrapper[4767]: I0317 16:03:53.073277 4767 generic.go:334] "Generic (PLEG): container finished" podID="d17172fd-996e-496e-92b5-9cd19edc65f4" containerID="8bf7818523bacace760493e89ff454ca91e19b8de2463beb219f503b023448e1" exitCode=2 Mar 17 16:03:53 crc kubenswrapper[4767]: I0317 16:03:53.073379 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17172fd-996e-496e-92b5-9cd19edc65f4","Type":"ContainerDied","Data":"8bf7818523bacace760493e89ff454ca91e19b8de2463beb219f503b023448e1"} Mar 17 16:03:53 crc kubenswrapper[4767]: I0317 16:03:53.082572 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf","Type":"ContainerStarted","Data":"c73c64275664e5084cbfcf1f83a76c8f702ab625945ed36e6b24bd1759ac09ec"} Mar 17 16:03:53 crc kubenswrapper[4767]: I0317 16:03:53.086447 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" event={"ID":"2f4160a8-203a-41e6-a90f-f0800ae776b1","Type":"ContainerStarted","Data":"89469eed3c5666995f50f5309cf2b253b49b11c04e7890fe945935f52393344d"} Mar 17 16:03:53 crc kubenswrapper[4767]: I0317 16:03:53.090529 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7874c48754-xp7vx" Mar 17 16:03:53 crc kubenswrapper[4767]: I0317 16:03:53.090997 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-976fc" event={"ID":"25debbd1-41e4-42db-af74-c058c0d6d9b2","Type":"ContainerStarted","Data":"66f633e8bd912a5ee681ce4f080954b74553019edf6c5e56558a68eed3012220"} Mar 17 16:03:53 crc kubenswrapper[4767]: I0317 16:03:53.091092 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-976fc" event={"ID":"25debbd1-41e4-42db-af74-c058c0d6d9b2","Type":"ContainerStarted","Data":"0afa0e5e22c51c3249ffa5a28a8cdc4b351411b46284dfdcf89b0e0cb65a6a65"} Mar 17 16:03:53 crc kubenswrapper[4767]: I0317 16:03:53.733937 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7874c48754-xp7vx"] Mar 17 16:03:53 crc kubenswrapper[4767]: I0317 16:03:53.754791 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7874c48754-xp7vx"] Mar 17 16:03:54 crc kubenswrapper[4767]: I0317 16:03:54.113216 4767 generic.go:334] "Generic (PLEG): container finished" podID="2f4160a8-203a-41e6-a90f-f0800ae776b1" containerID="41fd51b01636d01492ed5e2e39508da2604df8f281a4571a0ad83f67237944a3" exitCode=0 Mar 17 16:03:54 crc kubenswrapper[4767]: I0317 16:03:54.113364 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" event={"ID":"2f4160a8-203a-41e6-a90f-f0800ae776b1","Type":"ContainerDied","Data":"41fd51b01636d01492ed5e2e39508da2604df8f281a4571a0ad83f67237944a3"} Mar 17 16:03:54 crc kubenswrapper[4767]: I0317 16:03:54.118156 4767 generic.go:334] "Generic (PLEG): container finished" podID="25debbd1-41e4-42db-af74-c058c0d6d9b2" containerID="66f633e8bd912a5ee681ce4f080954b74553019edf6c5e56558a68eed3012220" exitCode=0 Mar 17 16:03:54 crc kubenswrapper[4767]: I0317 16:03:54.118210 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-976fc" event={"ID":"25debbd1-41e4-42db-af74-c058c0d6d9b2","Type":"ContainerDied","Data":"66f633e8bd912a5ee681ce4f080954b74553019edf6c5e56558a68eed3012220"} Mar 17 16:03:55 crc kubenswrapper[4767]: I0317 16:03:55.148267 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-976fc" event={"ID":"25debbd1-41e4-42db-af74-c058c0d6d9b2","Type":"ContainerStarted","Data":"577a346deaf9c1ff373cbea505c9f9a9c43d604bae7e9e8a5561d4a95616dc78"} Mar 17 16:03:55 crc kubenswrapper[4767]: I0317 16:03:55.168118 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f964f8b0-adf7-4bf3-880c-f123640015af","Type":"ContainerStarted","Data":"0ac557932f4c3ff3227bd936c176e49d7d8b341f3367b206c3aafa0083f437c2"} Mar 17 16:03:55 crc kubenswrapper[4767]: I0317 16:03:55.373925 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" path="/var/lib/kubelet/pods/22f8c42f-134c-47ed-baa8-0f176a362d38/volumes" Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.483751 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf","Type":"ContainerStarted","Data":"32c69035be2db4a823c444722a55b1eba080d1ae8802712b43df538d830a9c40"} Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.484346 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf","Type":"ContainerStarted","Data":"5f6dc385e1ae9b037ed9b0027b71268eb9f3c9e4febb6b7b87b97ddb7bad6f1e"} Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.495429 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" event={"ID":"2f4160a8-203a-41e6-a90f-f0800ae776b1","Type":"ContainerStarted","Data":"7b426e927de42d1e528bab380f9497633286b54dc26035261cb9ceeff790bb07"} Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.495530 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.503918 4767 generic.go:334] "Generic (PLEG): container finished" podID="25debbd1-41e4-42db-af74-c058c0d6d9b2" containerID="577a346deaf9c1ff373cbea505c9f9a9c43d604bae7e9e8a5561d4a95616dc78" exitCode=0 Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.504029 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-976fc" event={"ID":"25debbd1-41e4-42db-af74-c058c0d6d9b2","Type":"ContainerDied","Data":"577a346deaf9c1ff373cbea505c9f9a9c43d604bae7e9e8a5561d4a95616dc78"} Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.504966 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.351451551 podStartE2EDuration="8.504947633s" podCreationTimestamp="2026-03-17 16:03:48 +0000 UTC" firstStartedPulling="2026-03-17 16:03:52.6574802 +0000 UTC m=+1624.070796247" lastFinishedPulling="2026-03-17 16:03:53.810976282 +0000 UTC m=+1625.224292329" observedRunningTime="2026-03-17 16:03:56.503027678 +0000 UTC m=+1627.916343725" watchObservedRunningTime="2026-03-17 16:03:56.504947633 +0000 UTC m=+1627.918263680" Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.524619 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f964f8b0-adf7-4bf3-880c-f123640015af","Type":"ContainerStarted","Data":"83ed91c2f64239212c25fb176a292cc5eac0e8a1e50a4e782d411aab9ff7a24e"} Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.524883 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f964f8b0-adf7-4bf3-880c-f123640015af" containerName="cinder-api-log" containerID="cri-o://0ac557932f4c3ff3227bd936c176e49d7d8b341f3367b206c3aafa0083f437c2" gracePeriod=30 Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.525044 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.525186 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f964f8b0-adf7-4bf3-880c-f123640015af" containerName="cinder-api" containerID="cri-o://83ed91c2f64239212c25fb176a292cc5eac0e8a1e50a4e782d411aab9ff7a24e" gracePeriod=30 Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.616590 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" podStartSLOduration=7.616561716 podStartE2EDuration="7.616561716s" podCreationTimestamp="2026-03-17 16:03:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:56.529047562 +0000 UTC m=+1627.942363619" watchObservedRunningTime="2026-03-17 16:03:56.616561716 +0000 UTC m=+1628.029877763" Mar 17 16:03:56 crc kubenswrapper[4767]: I0317 16:03:56.688616 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.688592006 podStartE2EDuration="7.688592006s" podCreationTimestamp="2026-03-17 16:03:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:03:56.565009181 +0000 UTC m=+1627.978325248" watchObservedRunningTime="2026-03-17 16:03:56.688592006 +0000 UTC m=+1628.101908053" Mar 17 16:03:57 crc kubenswrapper[4767]: I0317 16:03:57.265890 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:57 crc kubenswrapper[4767]: I0317 16:03:57.267489 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:03:57 crc kubenswrapper[4767]: I0317 16:03:57.590517 4767 generic.go:334] "Generic (PLEG): container finished" podID="f964f8b0-adf7-4bf3-880c-f123640015af" containerID="83ed91c2f64239212c25fb176a292cc5eac0e8a1e50a4e782d411aab9ff7a24e" exitCode=0 Mar 17 16:03:57 crc kubenswrapper[4767]: I0317 16:03:57.590843 4767 generic.go:334] "Generic (PLEG): container finished" podID="f964f8b0-adf7-4bf3-880c-f123640015af" containerID="0ac557932f4c3ff3227bd936c176e49d7d8b341f3367b206c3aafa0083f437c2" exitCode=143 Mar 17 16:03:57 crc kubenswrapper[4767]: I0317 16:03:57.590938 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f964f8b0-adf7-4bf3-880c-f123640015af","Type":"ContainerDied","Data":"83ed91c2f64239212c25fb176a292cc5eac0e8a1e50a4e782d411aab9ff7a24e"} Mar 17 16:03:57 crc kubenswrapper[4767]: I0317 16:03:57.590976 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f964f8b0-adf7-4bf3-880c-f123640015af","Type":"ContainerDied","Data":"0ac557932f4c3ff3227bd936c176e49d7d8b341f3367b206c3aafa0083f437c2"} Mar 17 16:03:57 crc kubenswrapper[4767]: I0317 16:03:57.617743 4767 generic.go:334] "Generic (PLEG): container finished" podID="d17172fd-996e-496e-92b5-9cd19edc65f4" containerID="94848df684c7d16e35f3425fecf44bad6bb45f6cba749b60673a0a0b7fe71222" exitCode=0 Mar 17 16:03:57 crc kubenswrapper[4767]: I0317 16:03:57.617957 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17172fd-996e-496e-92b5-9cd19edc65f4","Type":"ContainerDied","Data":"94848df684c7d16e35f3425fecf44bad6bb45f6cba749b60673a0a0b7fe71222"} Mar 17 16:03:57 crc kubenswrapper[4767]: I0317 16:03:57.977539 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.052380 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-scripts\") pod \"f964f8b0-adf7-4bf3-880c-f123640015af\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.052535 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data\") pod \"f964f8b0-adf7-4bf3-880c-f123640015af\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.052787 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f964f8b0-adf7-4bf3-880c-f123640015af-etc-machine-id\") pod \"f964f8b0-adf7-4bf3-880c-f123640015af\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.052837 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data-custom\") pod \"f964f8b0-adf7-4bf3-880c-f123640015af\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.052895 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-combined-ca-bundle\") pod \"f964f8b0-adf7-4bf3-880c-f123640015af\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.052940 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgnks\" (UniqueName: \"kubernetes.io/projected/f964f8b0-adf7-4bf3-880c-f123640015af-kube-api-access-dgnks\") pod \"f964f8b0-adf7-4bf3-880c-f123640015af\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.053012 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f964f8b0-adf7-4bf3-880c-f123640015af-logs\") pod \"f964f8b0-adf7-4bf3-880c-f123640015af\" (UID: \"f964f8b0-adf7-4bf3-880c-f123640015af\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.053205 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f964f8b0-adf7-4bf3-880c-f123640015af-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f964f8b0-adf7-4bf3-880c-f123640015af" (UID: "f964f8b0-adf7-4bf3-880c-f123640015af"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.055148 4767 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f964f8b0-adf7-4bf3-880c-f123640015af-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.066422 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f964f8b0-adf7-4bf3-880c-f123640015af-logs" (OuterVolumeSpecName: "logs") pod "f964f8b0-adf7-4bf3-880c-f123640015af" (UID: "f964f8b0-adf7-4bf3-880c-f123640015af"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.093284 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f964f8b0-adf7-4bf3-880c-f123640015af" (UID: "f964f8b0-adf7-4bf3-880c-f123640015af"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.109560 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-scripts" (OuterVolumeSpecName: "scripts") pod "f964f8b0-adf7-4bf3-880c-f123640015af" (UID: "f964f8b0-adf7-4bf3-880c-f123640015af"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.120622 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f964f8b0-adf7-4bf3-880c-f123640015af-kube-api-access-dgnks" (OuterVolumeSpecName: "kube-api-access-dgnks") pod "f964f8b0-adf7-4bf3-880c-f123640015af" (UID: "f964f8b0-adf7-4bf3-880c-f123640015af"). InnerVolumeSpecName "kube-api-access-dgnks". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.140961 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f964f8b0-adf7-4bf3-880c-f123640015af" (UID: "f964f8b0-adf7-4bf3-880c-f123640015af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.160657 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f964f8b0-adf7-4bf3-880c-f123640015af-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.160692 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.160707 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.160716 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.160728 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgnks\" (UniqueName: \"kubernetes.io/projected/f964f8b0-adf7-4bf3-880c-f123640015af-kube-api-access-dgnks\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.231285 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data" (OuterVolumeSpecName: "config-data") pod "f964f8b0-adf7-4bf3-880c-f123640015af" (UID: "f964f8b0-adf7-4bf3-880c-f123640015af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.263443 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f964f8b0-adf7-4bf3-880c-f123640015af-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.271099 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.336261 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.365710 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-config-data\") pod \"d17172fd-996e-496e-92b5-9cd19edc65f4\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.365777 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-run-httpd\") pod \"d17172fd-996e-496e-92b5-9cd19edc65f4\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.365893 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-combined-ca-bundle\") pod \"d17172fd-996e-496e-92b5-9cd19edc65f4\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.366069 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-sg-core-conf-yaml\") pod \"d17172fd-996e-496e-92b5-9cd19edc65f4\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.366104 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-scripts\") pod \"d17172fd-996e-496e-92b5-9cd19edc65f4\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.366136 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-log-httpd\") pod \"d17172fd-996e-496e-92b5-9cd19edc65f4\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.366326 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25njk\" (UniqueName: \"kubernetes.io/projected/d17172fd-996e-496e-92b5-9cd19edc65f4-kube-api-access-25njk\") pod \"d17172fd-996e-496e-92b5-9cd19edc65f4\" (UID: \"d17172fd-996e-496e-92b5-9cd19edc65f4\") " Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.366353 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d17172fd-996e-496e-92b5-9cd19edc65f4" (UID: "d17172fd-996e-496e-92b5-9cd19edc65f4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.366498 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d17172fd-996e-496e-92b5-9cd19edc65f4" (UID: "d17172fd-996e-496e-92b5-9cd19edc65f4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.367871 4767 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.368301 4767 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17172fd-996e-496e-92b5-9cd19edc65f4-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.373637 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d17172fd-996e-496e-92b5-9cd19edc65f4-kube-api-access-25njk" (OuterVolumeSpecName: "kube-api-access-25njk") pod "d17172fd-996e-496e-92b5-9cd19edc65f4" (UID: "d17172fd-996e-496e-92b5-9cd19edc65f4"). InnerVolumeSpecName "kube-api-access-25njk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.376908 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-scripts" (OuterVolumeSpecName: "scripts") pod "d17172fd-996e-496e-92b5-9cd19edc65f4" (UID: "d17172fd-996e-496e-92b5-9cd19edc65f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.403400 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d17172fd-996e-496e-92b5-9cd19edc65f4" (UID: "d17172fd-996e-496e-92b5-9cd19edc65f4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.421103 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-config-data" (OuterVolumeSpecName: "config-data") pod "d17172fd-996e-496e-92b5-9cd19edc65f4" (UID: "d17172fd-996e-496e-92b5-9cd19edc65f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.421465 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d17172fd-996e-496e-92b5-9cd19edc65f4" (UID: "d17172fd-996e-496e-92b5-9cd19edc65f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.470524 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.470571 4767 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.470582 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.470596 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25njk\" (UniqueName: \"kubernetes.io/projected/d17172fd-996e-496e-92b5-9cd19edc65f4-kube-api-access-25njk\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.470606 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17172fd-996e-496e-92b5-9cd19edc65f4-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.624589 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-79f98b8574-mc4pp" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.632277 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f964f8b0-adf7-4bf3-880c-f123640015af","Type":"ContainerDied","Data":"05f44bc81ba5e1bc2ac892fbd1f628a4ca9644b6c12c38a4614e16c3c8c84dec"} Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.632342 4767 scope.go:117] "RemoveContainer" containerID="83ed91c2f64239212c25fb176a292cc5eac0e8a1e50a4e782d411aab9ff7a24e" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.632535 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.641547 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17172fd-996e-496e-92b5-9cd19edc65f4","Type":"ContainerDied","Data":"b341537ccb4b136de4ee6b2f73d91bb94b0972196b2cc9404f3b92ace4936702"} Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.641655 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.647449 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-976fc" event={"ID":"25debbd1-41e4-42db-af74-c058c0d6d9b2","Type":"ContainerStarted","Data":"f36e327042d1a9ebde847b74eb2fed4597a8b076637f41fd9237463073f3bfb7"} Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.688432 4767 scope.go:117] "RemoveContainer" containerID="0ac557932f4c3ff3227bd936c176e49d7d8b341f3367b206c3aafa0083f437c2" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.701334 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-976fc" podStartSLOduration=10.850723349 podStartE2EDuration="14.701301612s" podCreationTimestamp="2026-03-17 16:03:44 +0000 UTC" firstStartedPulling="2026-03-17 16:03:53.387059397 +0000 UTC m=+1624.800375444" lastFinishedPulling="2026-03-17 16:03:57.23763766 +0000 UTC m=+1628.650953707" observedRunningTime="2026-03-17 16:03:58.694661892 +0000 UTC m=+1630.107977939" watchObservedRunningTime="2026-03-17 16:03:58.701301612 +0000 UTC m=+1630.114617659" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.730606 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8557564488-zkt7w" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.736159 4767 scope.go:117] "RemoveContainer" containerID="8bf7818523bacace760493e89ff454ca91e19b8de2463beb219f503b023448e1" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.744878 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.777332 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.816408 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 17 16:03:58 crc kubenswrapper[4767]: E0317 16:03:58.832879 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerName="barbican-api" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.832925 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerName="barbican-api" Mar 17 16:03:58 crc kubenswrapper[4767]: E0317 16:03:58.832954 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f964f8b0-adf7-4bf3-880c-f123640015af" containerName="cinder-api-log" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.832962 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f964f8b0-adf7-4bf3-880c-f123640015af" containerName="cinder-api-log" Mar 17 16:03:58 crc kubenswrapper[4767]: E0317 16:03:58.833003 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17172fd-996e-496e-92b5-9cd19edc65f4" containerName="sg-core" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.833010 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17172fd-996e-496e-92b5-9cd19edc65f4" containerName="sg-core" Mar 17 16:03:58 crc kubenswrapper[4767]: E0317 16:03:58.833029 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f964f8b0-adf7-4bf3-880c-f123640015af" containerName="cinder-api" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.833035 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f964f8b0-adf7-4bf3-880c-f123640015af" containerName="cinder-api" Mar 17 16:03:58 crc kubenswrapper[4767]: E0317 16:03:58.833052 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerName="barbican-api-log" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.833059 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerName="barbican-api-log" Mar 17 16:03:58 crc kubenswrapper[4767]: E0317 16:03:58.833078 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17172fd-996e-496e-92b5-9cd19edc65f4" containerName="ceilometer-notification-agent" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.833085 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17172fd-996e-496e-92b5-9cd19edc65f4" containerName="ceilometer-notification-agent" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.833563 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17172fd-996e-496e-92b5-9cd19edc65f4" containerName="ceilometer-notification-agent" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.833605 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerName="barbican-api" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.833626 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f964f8b0-adf7-4bf3-880c-f123640015af" containerName="cinder-api-log" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.833643 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="22f8c42f-134c-47ed-baa8-0f176a362d38" containerName="barbican-api-log" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.833661 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f964f8b0-adf7-4bf3-880c-f123640015af" containerName="cinder-api" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.833679 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17172fd-996e-496e-92b5-9cd19edc65f4" containerName="sg-core" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.836388 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.870481 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.871598 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.871812 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.872340 4767 scope.go:117] "RemoveContainer" containerID="94848df684c7d16e35f3425fecf44bad6bb45f6cba749b60673a0a0b7fe71222" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.911609 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.953888 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.966735 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.984082 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.985068 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.985152 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ca1d252-faff-4972-8ca6-5b66cf11de53-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.985453 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-scripts\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.985674 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ca1d252-faff-4972-8ca6-5b66cf11de53-logs\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.985754 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-config-data-custom\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.985804 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.985833 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jnld\" (UniqueName: \"kubernetes.io/projected/2ca1d252-faff-4972-8ca6-5b66cf11de53-kube-api-access-6jnld\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.985875 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-config-data\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.994573 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:03:58 crc kubenswrapper[4767]: I0317 16:03:58.998606 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.003183 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.003632 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.047348 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.068708 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-844b76bcb8-bt49d"] Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.069482 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-844b76bcb8-bt49d" podUID="7d518ac2-8f5c-4520-b3b4-8e01f6678974" containerName="placement-log" containerID="cri-o://6f3a2f449fb65e19cac86104789c3ad946517e7dd0bf835c4455fa34962b7df9" gracePeriod=30 Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.069675 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-844b76bcb8-bt49d" podUID="7d518ac2-8f5c-4520-b3b4-8e01f6678974" containerName="placement-api" containerID="cri-o://1f135306da0aaa11ef00c9f9f2c78a4ba01695e533b490b69523813a529507ce" gracePeriod=30 Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.089155 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-config-data-custom\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.089444 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.089539 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jnld\" (UniqueName: \"kubernetes.io/projected/2ca1d252-faff-4972-8ca6-5b66cf11de53-kube-api-access-6jnld\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.089583 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-config-data\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.089757 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.089861 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.089973 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ca1d252-faff-4972-8ca6-5b66cf11de53-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.090028 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-scripts\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.090214 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ca1d252-faff-4972-8ca6-5b66cf11de53-logs\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.094470 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ca1d252-faff-4972-8ca6-5b66cf11de53-logs\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.096373 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ca1d252-faff-4972-8ca6-5b66cf11de53-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.100796 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.112571 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.112817 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.114585 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-config-data-custom\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.119256 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jnld\" (UniqueName: \"kubernetes.io/projected/2ca1d252-faff-4972-8ca6-5b66cf11de53-kube-api-access-6jnld\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.119360 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-config-data\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.119797 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ca1d252-faff-4972-8ca6-5b66cf11de53-scripts\") pod \"cinder-api-0\" (UID: \"2ca1d252-faff-4972-8ca6-5b66cf11de53\") " pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.196460 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-config-data\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.196590 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-log-httpd\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.196893 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-scripts\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.196956 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.197074 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-run-httpd\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.197206 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.197282 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gzps\" (UniqueName: \"kubernetes.io/projected/0d2da299-4e48-4135-a749-7a2adf3f7ec9-kube-api-access-5gzps\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.216762 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.300409 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-config-data\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.300860 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-log-httpd\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.300938 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-scripts\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.300959 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.301025 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-run-httpd\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.301077 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.301111 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gzps\" (UniqueName: \"kubernetes.io/projected/0d2da299-4e48-4135-a749-7a2adf3f7ec9-kube-api-access-5gzps\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.304781 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-run-httpd\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.333381 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-log-httpd\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.487943 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-scripts\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.494364 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.494793 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-config-data\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.517445 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d17172fd-996e-496e-92b5-9cd19edc65f4" path="/var/lib/kubelet/pods/d17172fd-996e-496e-92b5-9cd19edc65f4/volumes" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.518484 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f964f8b0-adf7-4bf3-880c-f123640015af" path="/var/lib/kubelet/pods/f964f8b0-adf7-4bf3-880c-f123640015af/volumes" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.532225 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gzps\" (UniqueName: \"kubernetes.io/projected/0d2da299-4e48-4135-a749-7a2adf3f7ec9-kube-api-access-5gzps\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.532532 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.629143 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.771080 4767 generic.go:334] "Generic (PLEG): container finished" podID="7d518ac2-8f5c-4520-b3b4-8e01f6678974" containerID="6f3a2f449fb65e19cac86104789c3ad946517e7dd0bf835c4455fa34962b7df9" exitCode=143 Mar 17 16:03:59 crc kubenswrapper[4767]: I0317 16:03:59.771165 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-844b76bcb8-bt49d" event={"ID":"7d518ac2-8f5c-4520-b3b4-8e01f6678974","Type":"ContainerDied","Data":"6f3a2f449fb65e19cac86104789c3ad946517e7dd0bf835c4455fa34962b7df9"} Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.037436 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.160576 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562724-wqk5x"] Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.162820 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562724-wqk5x" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.175236 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.175533 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.175735 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.218268 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562724-wqk5x"] Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.248229 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.304775 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.543214 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqrmt\" (UniqueName: \"kubernetes.io/projected/4b1f5965-725f-47c1-bfc3-cdb8a3b7a696-kube-api-access-jqrmt\") pod \"auto-csr-approver-29562724-wqk5x\" (UID: \"4b1f5965-725f-47c1-bfc3-cdb8a3b7a696\") " pod="openshift-infra/auto-csr-approver-29562724-wqk5x" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.651197 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqrmt\" (UniqueName: \"kubernetes.io/projected/4b1f5965-725f-47c1-bfc3-cdb8a3b7a696-kube-api-access-jqrmt\") pod \"auto-csr-approver-29562724-wqk5x\" (UID: \"4b1f5965-725f-47c1-bfc3-cdb8a3b7a696\") " pod="openshift-infra/auto-csr-approver-29562724-wqk5x" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.708724 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sgpl5"] Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.709196 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" podUID="718f5a48-c119-4bae-9c9c-72ed544758ee" containerName="dnsmasq-dns" containerID="cri-o://7cbcbb684b7117829363a6d6637dc12889f3744a6903dce8fbfb35e6f52bd8f8" gracePeriod=10 Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.716440 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqrmt\" (UniqueName: \"kubernetes.io/projected/4b1f5965-725f-47c1-bfc3-cdb8a3b7a696-kube-api-access-jqrmt\") pod \"auto-csr-approver-29562724-wqk5x\" (UID: \"4b1f5965-725f-47c1-bfc3-cdb8a3b7a696\") " pod="openshift-infra/auto-csr-approver-29562724-wqk5x" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.729127 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.767843 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.823510 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" podUID="718f5a48-c119-4bae-9c9c-72ed544758ee" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.205:5353: connect: connection refused" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.864607 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562724-wqk5x" Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.889353 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ca1d252-faff-4972-8ca6-5b66cf11de53","Type":"ContainerStarted","Data":"2643d83d7ae99d069644c2981a427495e704960bddd9921ea6710cf540fec73a"} Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.929638 4767 generic.go:334] "Generic (PLEG): container finished" podID="718f5a48-c119-4bae-9c9c-72ed544758ee" containerID="7cbcbb684b7117829363a6d6637dc12889f3744a6903dce8fbfb35e6f52bd8f8" exitCode=0 Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.929803 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" event={"ID":"718f5a48-c119-4bae-9c9c-72ed544758ee","Type":"ContainerDied","Data":"7cbcbb684b7117829363a6d6637dc12889f3744a6903dce8fbfb35e6f52bd8f8"} Mar 17 16:04:00 crc kubenswrapper[4767]: I0317 16:04:00.949328 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d2da299-4e48-4135-a749-7a2adf3f7ec9","Type":"ContainerStarted","Data":"fbdf5f458101f27459fc0ac59386d618852caa139b35a854248efa096544ebfa"} Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.113651 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.689660 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.742940 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 17 16:04:01 crc kubenswrapper[4767]: E0317 16:04:01.743665 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="718f5a48-c119-4bae-9c9c-72ed544758ee" containerName="init" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.743686 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="718f5a48-c119-4bae-9c9c-72ed544758ee" containerName="init" Mar 17 16:04:01 crc kubenswrapper[4767]: E0317 16:04:01.743742 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="718f5a48-c119-4bae-9c9c-72ed544758ee" containerName="dnsmasq-dns" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.743750 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="718f5a48-c119-4bae-9c9c-72ed544758ee" containerName="dnsmasq-dns" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.744022 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="718f5a48-c119-4bae-9c9c-72ed544758ee" containerName="dnsmasq-dns" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.759050 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.769663 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-tcp7n" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.770089 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.770453 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.831072 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.860922 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-swift-storage-0\") pod \"718f5a48-c119-4bae-9c9c-72ed544758ee\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.861040 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-sb\") pod \"718f5a48-c119-4bae-9c9c-72ed544758ee\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.861125 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-svc\") pod \"718f5a48-c119-4bae-9c9c-72ed544758ee\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.861163 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-nb\") pod \"718f5a48-c119-4bae-9c9c-72ed544758ee\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.861785 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5v7x\" (UniqueName: \"kubernetes.io/projected/718f5a48-c119-4bae-9c9c-72ed544758ee-kube-api-access-q5v7x\") pod \"718f5a48-c119-4bae-9c9c-72ed544758ee\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.862148 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-config\") pod \"718f5a48-c119-4bae-9c9c-72ed544758ee\" (UID: \"718f5a48-c119-4bae-9c9c-72ed544758ee\") " Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.902316 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/718f5a48-c119-4bae-9c9c-72ed544758ee-kube-api-access-q5v7x" (OuterVolumeSpecName: "kube-api-access-q5v7x") pod "718f5a48-c119-4bae-9c9c-72ed544758ee" (UID: "718f5a48-c119-4bae-9c9c-72ed544758ee"). InnerVolumeSpecName "kube-api-access-q5v7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.967211 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-openstack-config\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.967259 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbldk\" (UniqueName: \"kubernetes.io/projected/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-kube-api-access-dbldk\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.967318 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-openstack-config-secret\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.967353 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.967412 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5v7x\" (UniqueName: \"kubernetes.io/projected/718f5a48-c119-4bae-9c9c-72ed544758ee-kube-api-access-q5v7x\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.983448 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562724-wqk5x"] Mar 17 16:04:01 crc kubenswrapper[4767]: I0317 16:04:01.985232 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562724-wqk5x" event={"ID":"4b1f5965-725f-47c1-bfc3-cdb8a3b7a696","Type":"ContainerStarted","Data":"98d99997248d2e3ab15c4f1a7f5f1e42acf4b7150e75b010e08de8d107d98c33"} Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.002273 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" event={"ID":"718f5a48-c119-4bae-9c9c-72ed544758ee","Type":"ContainerDied","Data":"54dec61b8f1f3cbb7d95d23d5348436aae4cab3a396a7a3433084e138d2d84f2"} Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.002360 4767 scope.go:117] "RemoveContainer" containerID="7cbcbb684b7117829363a6d6637dc12889f3744a6903dce8fbfb35e6f52bd8f8" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.002585 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-sgpl5" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.015298 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d2da299-4e48-4135-a749-7a2adf3f7ec9","Type":"ContainerStarted","Data":"b673dbe3c730b2a2acdf223e799163fe36598b87eb20e0ad865592a2c81c71ac"} Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.015717 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" containerName="cinder-scheduler" containerID="cri-o://5f6dc385e1ae9b037ed9b0027b71268eb9f3c9e4febb6b7b87b97ddb7bad6f1e" gracePeriod=30 Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.016722 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" containerName="probe" containerID="cri-o://32c69035be2db4a823c444722a55b1eba080d1ae8802712b43df538d830a9c40" gracePeriod=30 Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.072502 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-openstack-config\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.072825 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbldk\" (UniqueName: \"kubernetes.io/projected/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-kube-api-access-dbldk\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.073054 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-openstack-config-secret\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.073192 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.078136 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-openstack-config\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.091737 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbldk\" (UniqueName: \"kubernetes.io/projected/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-kube-api-access-dbldk\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.115021 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-openstack-config-secret\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.120721 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/049b9f3b-c6db-4274-9f45-c5cac0c0a17b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"049b9f3b-c6db-4274-9f45-c5cac0c0a17b\") " pod="openstack/openstackclient" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.352366 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "718f5a48-c119-4bae-9c9c-72ed544758ee" (UID: "718f5a48-c119-4bae-9c9c-72ed544758ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.362826 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-config" (OuterVolumeSpecName: "config") pod "718f5a48-c119-4bae-9c9c-72ed544758ee" (UID: "718f5a48-c119-4bae-9c9c-72ed544758ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.456310 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.456352 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.488710 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "718f5a48-c119-4bae-9c9c-72ed544758ee" (UID: "718f5a48-c119-4bae-9c9c-72ed544758ee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.559354 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.679523 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "718f5a48-c119-4bae-9c9c-72ed544758ee" (UID: "718f5a48-c119-4bae-9c9c-72ed544758ee"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.774236 4767 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.957392 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "718f5a48-c119-4bae-9c9c-72ed544758ee" (UID: "718f5a48-c119-4bae-9c9c-72ed544758ee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:02 crc kubenswrapper[4767]: I0317 16:04:02.990778 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/718f5a48-c119-4bae-9c9c-72ed544758ee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.062289 4767 generic.go:334] "Generic (PLEG): container finished" podID="7d518ac2-8f5c-4520-b3b4-8e01f6678974" containerID="1f135306da0aaa11ef00c9f9f2c78a4ba01695e533b490b69523813a529507ce" exitCode=0 Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.062601 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-844b76bcb8-bt49d" event={"ID":"7d518ac2-8f5c-4520-b3b4-8e01f6678974","Type":"ContainerDied","Data":"1f135306da0aaa11ef00c9f9f2c78a4ba01695e533b490b69523813a529507ce"} Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.082573 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ca1d252-faff-4972-8ca6-5b66cf11de53","Type":"ContainerStarted","Data":"f2c58ba5347dedba449af64977d386f76956660d385492c6b75f6387f02b3e97"} Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.401310 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.555890 4767 scope.go:117] "RemoveContainer" containerID="962e2a724c2df9cf43853d3b90406002a9710291d367f01a12a678eb7ca5b7a3" Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.627537 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sgpl5"] Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.643559 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-sgpl5"] Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.764240 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.876029 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-combined-ca-bundle\") pod \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.876131 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-config-data\") pod \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.876247 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d518ac2-8f5c-4520-b3b4-8e01f6678974-logs\") pod \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.876324 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-scripts\") pod \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.876496 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qflp\" (UniqueName: \"kubernetes.io/projected/7d518ac2-8f5c-4520-b3b4-8e01f6678974-kube-api-access-6qflp\") pod \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.876565 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-public-tls-certs\") pod \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.876729 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-internal-tls-certs\") pod \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\" (UID: \"7d518ac2-8f5c-4520-b3b4-8e01f6678974\") " Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.878756 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d518ac2-8f5c-4520-b3b4-8e01f6678974-logs" (OuterVolumeSpecName: "logs") pod "7d518ac2-8f5c-4520-b3b4-8e01f6678974" (UID: "7d518ac2-8f5c-4520-b3b4-8e01f6678974"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.884141 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d518ac2-8f5c-4520-b3b4-8e01f6678974-kube-api-access-6qflp" (OuterVolumeSpecName: "kube-api-access-6qflp") pod "7d518ac2-8f5c-4520-b3b4-8e01f6678974" (UID: "7d518ac2-8f5c-4520-b3b4-8e01f6678974"). InnerVolumeSpecName "kube-api-access-6qflp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.885086 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-scripts" (OuterVolumeSpecName: "scripts") pod "7d518ac2-8f5c-4520-b3b4-8e01f6678974" (UID: "7d518ac2-8f5c-4520-b3b4-8e01f6678974"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.984359 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d518ac2-8f5c-4520-b3b4-8e01f6678974-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.984670 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:03 crc kubenswrapper[4767]: I0317 16:04:03.984683 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qflp\" (UniqueName: \"kubernetes.io/projected/7d518ac2-8f5c-4520-b3b4-8e01f6678974-kube-api-access-6qflp\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.087458 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d518ac2-8f5c-4520-b3b4-8e01f6678974" (UID: "7d518ac2-8f5c-4520-b3b4-8e01f6678974"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.154940 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-config-data" (OuterVolumeSpecName: "config-data") pod "7d518ac2-8f5c-4520-b3b4-8e01f6678974" (UID: "7d518ac2-8f5c-4520-b3b4-8e01f6678974"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.167653 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.167732 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.167811 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.169004 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.169072 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" gracePeriod=600 Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.180623 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d2da299-4e48-4135-a749-7a2adf3f7ec9","Type":"ContainerStarted","Data":"dfe0080987b662004c5cc0a001db82ea1328fb2304d1b10898150fe5a5f55213"} Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.199141 4767 generic.go:334] "Generic (PLEG): container finished" podID="5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" containerID="32c69035be2db4a823c444722a55b1eba080d1ae8802712b43df538d830a9c40" exitCode=0 Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.199218 4767 generic.go:334] "Generic (PLEG): container finished" podID="5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" containerID="5f6dc385e1ae9b037ed9b0027b71268eb9f3c9e4febb6b7b87b97ddb7bad6f1e" exitCode=0 Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.199302 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.199314 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf","Type":"ContainerDied","Data":"32c69035be2db4a823c444722a55b1eba080d1ae8802712b43df538d830a9c40"} Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.199352 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf","Type":"ContainerDied","Data":"5f6dc385e1ae9b037ed9b0027b71268eb9f3c9e4febb6b7b87b97ddb7bad6f1e"} Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.199323 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.280317 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-844b76bcb8-bt49d" event={"ID":"7d518ac2-8f5c-4520-b3b4-8e01f6678974","Type":"ContainerDied","Data":"6ce5194d8010af588fcab7c32752c462ca0e06ce8137a51c00273edaadf3c9bf"} Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.280392 4767 scope.go:117] "RemoveContainer" containerID="1f135306da0aaa11ef00c9f9f2c78a4ba01695e533b490b69523813a529507ce" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.280585 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-844b76bcb8-bt49d" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.286809 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7d518ac2-8f5c-4520-b3b4-8e01f6678974" (UID: "7d518ac2-8f5c-4520-b3b4-8e01f6678974"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.301792 4767 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.310089 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7d518ac2-8f5c-4520-b3b4-8e01f6678974" (UID: "7d518ac2-8f5c-4520-b3b4-8e01f6678974"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:04 crc kubenswrapper[4767]: E0317 16:04:04.332245 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.404314 4767 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d518ac2-8f5c-4520-b3b4-8e01f6678974-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.433696 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.468860 4767 scope.go:117] "RemoveContainer" containerID="6f3a2f449fb65e19cac86104789c3ad946517e7dd0bf835c4455fa34962b7df9" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.483283 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-64d5cdf6bf-rvspf" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.529959 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-976fc" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.530466 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-976fc" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.598339 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-667f69cbdb-5qvs7"] Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.598725 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-667f69cbdb-5qvs7" podUID="86426b6e-ec55-49d4-804e-75b204239fa1" containerName="neutron-api" containerID="cri-o://c2b9c871dd75268d8313f0a8947b962e08fc00dd9b8852847187eaeebe41cf7d" gracePeriod=30 Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.599190 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-667f69cbdb-5qvs7" podUID="86426b6e-ec55-49d4-804e-75b204239fa1" containerName="neutron-httpd" containerID="cri-o://d8b69c7107b0c5f91af953d758844dc3ee89532cd4ca4ee00b6b4829ad5f3f18" gracePeriod=30 Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.622848 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-976fc" Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.675384 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-844b76bcb8-bt49d"] Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.699069 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-844b76bcb8-bt49d"] Mar 17 16:04:04 crc kubenswrapper[4767]: I0317 16:04:04.969219 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.032718 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-etc-machine-id\") pod \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.033379 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data\") pod \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.033465 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-scripts\") pod \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.033577 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-combined-ca-bundle\") pod \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.033700 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb6xw\" (UniqueName: \"kubernetes.io/projected/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-kube-api-access-zb6xw\") pod \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.034021 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data-custom\") pod \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\" (UID: \"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf\") " Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.047375 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" (UID: "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.078958 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-kube-api-access-zb6xw" (OuterVolumeSpecName: "kube-api-access-zb6xw") pod "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" (UID: "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf"). InnerVolumeSpecName "kube-api-access-zb6xw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.079428 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" (UID: "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.137391 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.137424 4767 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.137434 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb6xw\" (UniqueName: \"kubernetes.io/projected/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-kube-api-access-zb6xw\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.147423 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-scripts" (OuterVolumeSpecName: "scripts") pod "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" (UID: "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.278136 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.278869 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" (UID: "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.380541 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.397580 4767 generic.go:334] "Generic (PLEG): container finished" podID="86426b6e-ec55-49d4-804e-75b204239fa1" containerID="d8b69c7107b0c5f91af953d758844dc3ee89532cd4ca4ee00b6b4829ad5f3f18" exitCode=0 Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.424811 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="718f5a48-c119-4bae-9c9c-72ed544758ee" path="/var/lib/kubelet/pods/718f5a48-c119-4bae-9c9c-72ed544758ee/volumes" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.431625 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d518ac2-8f5c-4520-b3b4-8e01f6678974" path="/var/lib/kubelet/pods/7d518ac2-8f5c-4520-b3b4-8e01f6678974/volumes" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.433615 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.442810 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-667f69cbdb-5qvs7" event={"ID":"86426b6e-ec55-49d4-804e-75b204239fa1","Type":"ContainerDied","Data":"d8b69c7107b0c5f91af953d758844dc3ee89532cd4ca4ee00b6b4829ad5f3f18"} Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.442878 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5b4676bc-09a3-4b2d-9bc9-579b497bbfaf","Type":"ContainerDied","Data":"c73c64275664e5084cbfcf1f83a76c8f702ab625945ed36e6b24bd1759ac09ec"} Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.442908 4767 scope.go:117] "RemoveContainer" containerID="32c69035be2db4a823c444722a55b1eba080d1ae8802712b43df538d830a9c40" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.471823 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" exitCode=0 Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.471905 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c"} Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.472856 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:04:05 crc kubenswrapper[4767]: E0317 16:04:05.473161 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.478636 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data" (OuterVolumeSpecName: "config-data") pod "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" (UID: "5b4676bc-09a3-4b2d-9bc9-579b497bbfaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.480319 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"049b9f3b-c6db-4274-9f45-c5cac0c0a17b","Type":"ContainerStarted","Data":"5a30a8710203aa83db8ab989bfea7fb9bec639c4f4ce3dda4954832beb1582e5"} Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.483813 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.502753 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ca1d252-faff-4972-8ca6-5b66cf11de53","Type":"ContainerStarted","Data":"c38c96597e88759930d71205530782a5cc62da35f54bde621fc5ec6542abde51"} Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.513867 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.525548 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562724-wqk5x" event={"ID":"4b1f5965-725f-47c1-bfc3-cdb8a3b7a696","Type":"ContainerStarted","Data":"b1a0f869b4fcd37a4835c9b4c678d18fce3210feb179b919f81eef27e69fce14"} Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.582114 4767 scope.go:117] "RemoveContainer" containerID="5f6dc385e1ae9b037ed9b0027b71268eb9f3c9e4febb6b7b87b97ddb7bad6f1e" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.593281 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.5932446030000005 podStartE2EDuration="7.593244603s" podCreationTimestamp="2026-03-17 16:03:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:05.543214792 +0000 UTC m=+1636.956530859" watchObservedRunningTime="2026-03-17 16:04:05.593244603 +0000 UTC m=+1637.006560650" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.614049 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562724-wqk5x" podStartSLOduration=3.9728288259999998 podStartE2EDuration="5.614021297s" podCreationTimestamp="2026-03-17 16:04:00 +0000 UTC" firstStartedPulling="2026-03-17 16:04:01.915085152 +0000 UTC m=+1633.328401199" lastFinishedPulling="2026-03-17 16:04:03.556277633 +0000 UTC m=+1634.969593670" observedRunningTime="2026-03-17 16:04:05.576768712 +0000 UTC m=+1636.990084759" watchObservedRunningTime="2026-03-17 16:04:05.614021297 +0000 UTC m=+1637.027337344" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.651681 4767 scope.go:117] "RemoveContainer" containerID="53b1541419163e416532d56412b5286cd823f9a0b2d0b4b97086b6642dba72af" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.689592 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-976fc" Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.823817 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-976fc"] Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.904091 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 16:04:05 crc kubenswrapper[4767]: I0317 16:04:05.955357 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.036149 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 16:04:06 crc kubenswrapper[4767]: E0317 16:04:06.038045 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" containerName="probe" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.038064 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" containerName="probe" Mar 17 16:04:06 crc kubenswrapper[4767]: E0317 16:04:06.038087 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" containerName="cinder-scheduler" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.038096 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" containerName="cinder-scheduler" Mar 17 16:04:06 crc kubenswrapper[4767]: E0317 16:04:06.038131 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d518ac2-8f5c-4520-b3b4-8e01f6678974" containerName="placement-log" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.038138 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d518ac2-8f5c-4520-b3b4-8e01f6678974" containerName="placement-log" Mar 17 16:04:06 crc kubenswrapper[4767]: E0317 16:04:06.038196 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d518ac2-8f5c-4520-b3b4-8e01f6678974" containerName="placement-api" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.038203 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d518ac2-8f5c-4520-b3b4-8e01f6678974" containerName="placement-api" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.038715 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d518ac2-8f5c-4520-b3b4-8e01f6678974" containerName="placement-api" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.038756 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d518ac2-8f5c-4520-b3b4-8e01f6678974" containerName="placement-log" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.038783 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" containerName="probe" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.038792 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" containerName="cinder-scheduler" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.042229 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.046443 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.150289 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.159571 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.159832 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f03669c-8635-4948-b89c-0f2c7de39718-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.159905 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.159952 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-scripts\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.160020 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ggqn\" (UniqueName: \"kubernetes.io/projected/7f03669c-8635-4948-b89c-0f2c7de39718-kube-api-access-2ggqn\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.160076 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-config-data\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.263080 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-scripts\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.263213 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ggqn\" (UniqueName: \"kubernetes.io/projected/7f03669c-8635-4948-b89c-0f2c7de39718-kube-api-access-2ggqn\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.263245 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-config-data\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.263286 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.263392 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f03669c-8635-4948-b89c-0f2c7de39718-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.263455 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.264777 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f03669c-8635-4948-b89c-0f2c7de39718-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.270411 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-config-data\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.272726 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-scripts\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.273768 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.292874 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f03669c-8635-4948-b89c-0f2c7de39718-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.297764 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ggqn\" (UniqueName: \"kubernetes.io/projected/7f03669c-8635-4948-b89c-0f2c7de39718-kube-api-access-2ggqn\") pod \"cinder-scheduler-0\" (UID: \"7f03669c-8635-4948-b89c-0f2c7de39718\") " pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.443807 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-67c55f74cc-tpfc4" podUID="16264c79-b41d-43e1-a692-e084ae52e928" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.200:9696/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.444938 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.611532 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d2da299-4e48-4135-a749-7a2adf3f7ec9","Type":"ContainerStarted","Data":"e2a40d99e8f68af20b6fa59961e265d0c3832a7052b6c49304e606de6ba4d2da"} Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.654587 4767 generic.go:334] "Generic (PLEG): container finished" podID="4b1f5965-725f-47c1-bfc3-cdb8a3b7a696" containerID="b1a0f869b4fcd37a4835c9b4c678d18fce3210feb179b919f81eef27e69fce14" exitCode=0 Mar 17 16:04:06 crc kubenswrapper[4767]: I0317 16:04:06.654958 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562724-wqk5x" event={"ID":"4b1f5965-725f-47c1-bfc3-cdb8a3b7a696","Type":"ContainerDied","Data":"b1a0f869b4fcd37a4835c9b4c678d18fce3210feb179b919f81eef27e69fce14"} Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.208767 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.351499 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-crqkh"] Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.354683 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.408265 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b4676bc-09a3-4b2d-9bc9-579b497bbfaf" path="/var/lib/kubelet/pods/5b4676bc-09a3-4b2d-9bc9-579b497bbfaf/volumes" Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.410716 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-crqkh"] Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.518895 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-utilities\") pod \"redhat-marketplace-crqkh\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.519032 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-catalog-content\") pod \"redhat-marketplace-crqkh\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.519101 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckvrd\" (UniqueName: \"kubernetes.io/projected/4ad1c7b7-377f-488c-af38-bf30a79c3589-kube-api-access-ckvrd\") pod \"redhat-marketplace-crqkh\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.632915 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-utilities\") pod \"redhat-marketplace-crqkh\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.633001 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-catalog-content\") pod \"redhat-marketplace-crqkh\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.633034 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckvrd\" (UniqueName: \"kubernetes.io/projected/4ad1c7b7-377f-488c-af38-bf30a79c3589-kube-api-access-ckvrd\") pod \"redhat-marketplace-crqkh\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.633574 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-utilities\") pod \"redhat-marketplace-crqkh\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.633804 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-catalog-content\") pod \"redhat-marketplace-crqkh\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.667829 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckvrd\" (UniqueName: \"kubernetes.io/projected/4ad1c7b7-377f-488c-af38-bf30a79c3589-kube-api-access-ckvrd\") pod \"redhat-marketplace-crqkh\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.673226 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7f03669c-8635-4948-b89c-0f2c7de39718","Type":"ContainerStarted","Data":"5c46833942ba8a5a5658c8603365a862e30e77e91444d66c0142f17bee673432"} Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.676700 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-976fc" podUID="25debbd1-41e4-42db-af74-c058c0d6d9b2" containerName="registry-server" containerID="cri-o://f36e327042d1a9ebde847b74eb2fed4597a8b076637f41fd9237463073f3bfb7" gracePeriod=2 Mar 17 16:04:07 crc kubenswrapper[4767]: I0317 16:04:07.708945 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:08 crc kubenswrapper[4767]: I0317 16:04:08.874945 4767 generic.go:334] "Generic (PLEG): container finished" podID="25debbd1-41e4-42db-af74-c058c0d6d9b2" containerID="f36e327042d1a9ebde847b74eb2fed4597a8b076637f41fd9237463073f3bfb7" exitCode=0 Mar 17 16:04:08 crc kubenswrapper[4767]: I0317 16:04:08.875578 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-976fc" event={"ID":"25debbd1-41e4-42db-af74-c058c0d6d9b2","Type":"ContainerDied","Data":"f36e327042d1a9ebde847b74eb2fed4597a8b076637f41fd9237463073f3bfb7"} Mar 17 16:04:08 crc kubenswrapper[4767]: I0317 16:04:08.910451 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7f03669c-8635-4948-b89c-0f2c7de39718","Type":"ContainerStarted","Data":"72c392e8dd7e292a16bc4cfcbd28413bffec2553e5bfca74eaf17003de7080e8"} Mar 17 16:04:08 crc kubenswrapper[4767]: I0317 16:04:08.990581 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562724-wqk5x" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.047994 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqrmt\" (UniqueName: \"kubernetes.io/projected/4b1f5965-725f-47c1-bfc3-cdb8a3b7a696-kube-api-access-jqrmt\") pod \"4b1f5965-725f-47c1-bfc3-cdb8a3b7a696\" (UID: \"4b1f5965-725f-47c1-bfc3-cdb8a3b7a696\") " Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.057828 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b1f5965-725f-47c1-bfc3-cdb8a3b7a696-kube-api-access-jqrmt" (OuterVolumeSpecName: "kube-api-access-jqrmt") pod "4b1f5965-725f-47c1-bfc3-cdb8a3b7a696" (UID: "4b1f5965-725f-47c1-bfc3-cdb8a3b7a696"). InnerVolumeSpecName "kube-api-access-jqrmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.101714 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-crqkh"] Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.151622 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqrmt\" (UniqueName: \"kubernetes.io/projected/4b1f5965-725f-47c1-bfc3-cdb8a3b7a696-kube-api-access-jqrmt\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.219333 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-976fc" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.368719 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5nz6\" (UniqueName: \"kubernetes.io/projected/25debbd1-41e4-42db-af74-c058c0d6d9b2-kube-api-access-m5nz6\") pod \"25debbd1-41e4-42db-af74-c058c0d6d9b2\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.369058 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-utilities\") pod \"25debbd1-41e4-42db-af74-c058c0d6d9b2\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.369177 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-catalog-content\") pod \"25debbd1-41e4-42db-af74-c058c0d6d9b2\" (UID: \"25debbd1-41e4-42db-af74-c058c0d6d9b2\") " Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.374822 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-utilities" (OuterVolumeSpecName: "utilities") pod "25debbd1-41e4-42db-af74-c058c0d6d9b2" (UID: "25debbd1-41e4-42db-af74-c058c0d6d9b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.420860 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25debbd1-41e4-42db-af74-c058c0d6d9b2-kube-api-access-m5nz6" (OuterVolumeSpecName: "kube-api-access-m5nz6") pod "25debbd1-41e4-42db-af74-c058c0d6d9b2" (UID: "25debbd1-41e4-42db-af74-c058c0d6d9b2"). InnerVolumeSpecName "kube-api-access-m5nz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.509944 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.513445 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5nz6\" (UniqueName: \"kubernetes.io/projected/25debbd1-41e4-42db-af74-c058c0d6d9b2-kube-api-access-m5nz6\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.627406 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25debbd1-41e4-42db-af74-c058c0d6d9b2" (UID: "25debbd1-41e4-42db-af74-c058c0d6d9b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.721534 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25debbd1-41e4-42db-af74-c058c0d6d9b2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.955186 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562724-wqk5x" event={"ID":"4b1f5965-725f-47c1-bfc3-cdb8a3b7a696","Type":"ContainerDied","Data":"98d99997248d2e3ab15c4f1a7f5f1e42acf4b7150e75b010e08de8d107d98c33"} Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.955505 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98d99997248d2e3ab15c4f1a7f5f1e42acf4b7150e75b010e08de8d107d98c33" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.955511 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562724-wqk5x" Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.971558 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7f03669c-8635-4948-b89c-0f2c7de39718","Type":"ContainerStarted","Data":"d2bddddcf72e250e5875d484b0ca57b83aa2a7bcb91fb820de0df542cea29f65"} Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.976515 4767 generic.go:334] "Generic (PLEG): container finished" podID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerID="f2a60414104af01ed05aeb1eb576cad1646f14e32efe4085e03d6c5de1aefcdb" exitCode=0 Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.976596 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crqkh" event={"ID":"4ad1c7b7-377f-488c-af38-bf30a79c3589","Type":"ContainerDied","Data":"f2a60414104af01ed05aeb1eb576cad1646f14e32efe4085e03d6c5de1aefcdb"} Mar 17 16:04:09 crc kubenswrapper[4767]: I0317 16:04:09.976630 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crqkh" event={"ID":"4ad1c7b7-377f-488c-af38-bf30a79c3589","Type":"ContainerStarted","Data":"ddc2cee446a82701e0054130b4a12a4a20bbf6e195e503d9fde5759efdb3b2e8"} Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.029562 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.029529728 podStartE2EDuration="5.029529728s" podCreationTimestamp="2026-03-17 16:04:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:10.020775557 +0000 UTC m=+1641.434091614" watchObservedRunningTime="2026-03-17 16:04:10.029529728 +0000 UTC m=+1641.442845775" Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.031281 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d2da299-4e48-4135-a749-7a2adf3f7ec9","Type":"ContainerStarted","Data":"b4763a53b116968d26349a0361954feba8bb8491eb1e171954857b214c937eb0"} Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.032759 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 16:04:10 crc kubenswrapper[4767]: E0317 16:04:10.051441 4767 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d518ac2_8f5c_4520_b3b4_8e01f6678974.slice/crio-6ce5194d8010af588fcab7c32752c462ca0e06ce8137a51c00273edaadf3c9bf: Error finding container 6ce5194d8010af588fcab7c32752c462ca0e06ce8137a51c00273edaadf3c9bf: Status 404 returned error can't find the container with id 6ce5194d8010af588fcab7c32752c462ca0e06ce8137a51c00273edaadf3c9bf Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.053045 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf964f8b0_adf7_4bf3_880c_f123640015af.slice/crio-conmon-0ac557932f4c3ff3227bd936c176e49d7d8b341f3367b206c3aafa0083f437c2.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf964f8b0_adf7_4bf3_880c_f123640015af.slice/crio-conmon-0ac557932f4c3ff3227bd936c176e49d7d8b341f3367b206c3aafa0083f437c2.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.053118 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf964f8b0_adf7_4bf3_880c_f123640015af.slice/crio-0ac557932f4c3ff3227bd936c176e49d7d8b341f3367b206c3aafa0083f437c2.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf964f8b0_adf7_4bf3_880c_f123640015af.slice/crio-0ac557932f4c3ff3227bd936c176e49d7d8b341f3367b206c3aafa0083f437c2.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.053262 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25debbd1_41e4_42db_af74_c058c0d6d9b2.slice/crio-conmon-577a346deaf9c1ff373cbea505c9f9a9c43d604bae7e9e8a5561d4a95616dc78.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25debbd1_41e4_42db_af74_c058c0d6d9b2.slice/crio-conmon-577a346deaf9c1ff373cbea505c9f9a9c43d604bae7e9e8a5561d4a95616dc78.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.053320 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4676bc_09a3_4b2d_9bc9_579b497bbfaf.slice/crio-conmon-5f6dc385e1ae9b037ed9b0027b71268eb9f3c9e4febb6b7b87b97ddb7bad6f1e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4676bc_09a3_4b2d_9bc9_579b497bbfaf.slice/crio-conmon-5f6dc385e1ae9b037ed9b0027b71268eb9f3c9e4febb6b7b87b97ddb7bad6f1e.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.053532 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25debbd1_41e4_42db_af74_c058c0d6d9b2.slice/crio-577a346deaf9c1ff373cbea505c9f9a9c43d604bae7e9e8a5561d4a95616dc78.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25debbd1_41e4_42db_af74_c058c0d6d9b2.slice/crio-577a346deaf9c1ff373cbea505c9f9a9c43d604bae7e9e8a5561d4a95616dc78.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.077444 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4676bc_09a3_4b2d_9bc9_579b497bbfaf.slice/crio-5f6dc385e1ae9b037ed9b0027b71268eb9f3c9e4febb6b7b87b97ddb7bad6f1e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4676bc_09a3_4b2d_9bc9_579b497bbfaf.slice/crio-5f6dc385e1ae9b037ed9b0027b71268eb9f3c9e4febb6b7b87b97ddb7bad6f1e.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.077513 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf964f8b0_adf7_4bf3_880c_f123640015af.slice/crio-conmon-83ed91c2f64239212c25fb176a292cc5eac0e8a1e50a4e782d411aab9ff7a24e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf964f8b0_adf7_4bf3_880c_f123640015af.slice/crio-conmon-83ed91c2f64239212c25fb176a292cc5eac0e8a1e50a4e782d411aab9ff7a24e.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.077532 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf964f8b0_adf7_4bf3_880c_f123640015af.slice/crio-83ed91c2f64239212c25fb176a292cc5eac0e8a1e50a4e782d411aab9ff7a24e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf964f8b0_adf7_4bf3_880c_f123640015af.slice/crio-83ed91c2f64239212c25fb176a292cc5eac0e8a1e50a4e782d411aab9ff7a24e.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.077557 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4676bc_09a3_4b2d_9bc9_579b497bbfaf.slice/crio-conmon-32c69035be2db4a823c444722a55b1eba080d1ae8802712b43df538d830a9c40.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4676bc_09a3_4b2d_9bc9_579b497bbfaf.slice/crio-conmon-32c69035be2db4a823c444722a55b1eba080d1ae8802712b43df538d830a9c40.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.077573 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4676bc_09a3_4b2d_9bc9_579b497bbfaf.slice/crio-32c69035be2db4a823c444722a55b1eba080d1ae8802712b43df538d830a9c40.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4676bc_09a3_4b2d_9bc9_579b497bbfaf.slice/crio-32c69035be2db4a823c444722a55b1eba080d1ae8802712b43df538d830a9c40.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.077671 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25debbd1_41e4_42db_af74_c058c0d6d9b2.slice/crio-conmon-f36e327042d1a9ebde847b74eb2fed4597a8b076637f41fd9237463073f3bfb7.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25debbd1_41e4_42db_af74_c058c0d6d9b2.slice/crio-conmon-f36e327042d1a9ebde847b74eb2fed4597a8b076637f41fd9237463073f3bfb7.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.077703 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25debbd1_41e4_42db_af74_c058c0d6d9b2.slice/crio-f36e327042d1a9ebde847b74eb2fed4597a8b076637f41fd9237463073f3bfb7.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25debbd1_41e4_42db_af74_c058c0d6d9b2.slice/crio-f36e327042d1a9ebde847b74eb2fed4597a8b076637f41fd9237463073f3bfb7.scope: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.073178 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf964f8b0_adf7_4bf3_880c_f123640015af.slice/crio-05f44bc81ba5e1bc2ac892fbd1f628a4ca9644b6c12c38a4614e16c3c8c84dec WatchSource:0}: Error finding container 05f44bc81ba5e1bc2ac892fbd1f628a4ca9644b6c12c38a4614e16c3c8c84dec: Status 404 returned error can't find the container with id 05f44bc81ba5e1bc2ac892fbd1f628a4ca9644b6c12c38a4614e16c3c8c84dec Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.091071 4767 generic.go:334] "Generic (PLEG): container finished" podID="86426b6e-ec55-49d4-804e-75b204239fa1" containerID="c2b9c871dd75268d8313f0a8947b962e08fc00dd9b8852847187eaeebe41cf7d" exitCode=0 Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.091224 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-667f69cbdb-5qvs7" event={"ID":"86426b6e-ec55-49d4-804e-75b204239fa1","Type":"ContainerDied","Data":"c2b9c871dd75268d8313f0a8947b962e08fc00dd9b8852847187eaeebe41cf7d"} Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.120402 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4676bc_09a3_4b2d_9bc9_579b497bbfaf.slice/crio-c73c64275664e5084cbfcf1f83a76c8f702ab625945ed36e6b24bd1759ac09ec WatchSource:0}: Error finding container c73c64275664e5084cbfcf1f83a76c8f702ab625945ed36e6b24bd1759ac09ec: Status 404 returned error can't find the container with id c73c64275664e5084cbfcf1f83a76c8f702ab625945ed36e6b24bd1759ac09ec Mar 17 16:04:10 crc kubenswrapper[4767]: E0317 16:04:10.122532 4767 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: , extraDiskErr: could not stat "/var/log/pods/openshift-marketplace_community-operators-976fc_25debbd1-41e4-42db-af74-c058c0d6d9b2/extract-utilities/0.log" to get inode usage: stat /var/log/pods/openshift-marketplace_community-operators-976fc_25debbd1-41e4-42db-af74-c058c0d6d9b2/extract-utilities/0.log: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: W0317 16:04:10.137402 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b1f5965_725f_47c1_bfc3_cdb8a3b7a696.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b1f5965_725f_47c1_bfc3_cdb8a3b7a696.slice: no such file or directory Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.168809 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-976fc" event={"ID":"25debbd1-41e4-42db-af74-c058c0d6d9b2","Type":"ContainerDied","Data":"0afa0e5e22c51c3249ffa5a28a8cdc4b351411b46284dfdcf89b0e0cb65a6a65"} Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.168885 4767 scope.go:117] "RemoveContainer" containerID="f36e327042d1a9ebde847b74eb2fed4597a8b076637f41fd9237463073f3bfb7" Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.169172 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-976fc" Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.204493 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.869778407 podStartE2EDuration="12.204466831s" podCreationTimestamp="2026-03-17 16:03:58 +0000 UTC" firstStartedPulling="2026-03-17 16:04:00.695990644 +0000 UTC m=+1632.109306691" lastFinishedPulling="2026-03-17 16:04:08.030679068 +0000 UTC m=+1639.443995115" observedRunningTime="2026-03-17 16:04:10.098213872 +0000 UTC m=+1641.511529929" watchObservedRunningTime="2026-03-17 16:04:10.204466831 +0000 UTC m=+1641.617782878" Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.283199 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562718-28mzx"] Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.326744 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562718-28mzx"] Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.519451 4767 scope.go:117] "RemoveContainer" containerID="577a346deaf9c1ff373cbea505c9f9a9c43d604bae7e9e8a5561d4a95616dc78" Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.598426 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-976fc"] Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.737342 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-976fc"] Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.761807 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:04:10 crc kubenswrapper[4767]: E0317 16:04:10.763953 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod718f5a48_c119_4bae_9c9c_72ed544758ee.slice/crio-54dec61b8f1f3cbb7d95d23d5348436aae4cab3a396a7a3433084e138d2d84f2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd680dc75_619e_41ae_90cb_2f58898a33ac.slice/crio-conmon-0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d518ac2_8f5c_4520_b3b4_8e01f6678974.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc95699d_4853_4d4b_9239_019601fd7df8.slice/crio-conmon-1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff0f3ad0_1b55_4abb_9eb6_c07d82251bdd.slice/crio-conmon-2055be26e06e56bc2b2a7957901aafdce92a48fee3057028796f99ee80231978.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd17172fd_996e_496e_92b5_9cd19edc65f4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22f8c42f_134c_47ed_baa8_0f176a362d38.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4676bc_09a3_4b2d_9bc9_579b497bbfaf.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff0f3ad0_1b55_4abb_9eb6_c07d82251bdd.slice/crio-2055be26e06e56bc2b2a7957901aafdce92a48fee3057028796f99ee80231978.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd17172fd_996e_496e_92b5_9cd19edc65f4.slice/crio-8bf7818523bacace760493e89ff454ca91e19b8de2463beb219f503b023448e1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod718f5a48_c119_4bae_9c9c_72ed544758ee.slice/crio-7cbcbb684b7117829363a6d6637dc12889f3744a6903dce8fbfb35e6f52bd8f8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86426b6e_ec55_49d4_804e_75b204239fa1.slice/crio-conmon-d8b69c7107b0c5f91af953d758844dc3ee89532cd4ca4ee00b6b4829ad5f3f18.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22f8c42f_134c_47ed_baa8_0f176a362d38.slice/crio-fb68641df086d68c697275b20dc2997ea82074be958cf89d4c1031793e6d4fcb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod718f5a48_c119_4bae_9c9c_72ed544758ee.slice/crio-conmon-7cbcbb684b7117829363a6d6637dc12889f3744a6903dce8fbfb35e6f52bd8f8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc95699d_4853_4d4b_9239_019601fd7df8.slice/crio-1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd17172fd_996e_496e_92b5_9cd19edc65f4.slice/crio-conmon-8bf7818523bacace760493e89ff454ca91e19b8de2463beb219f503b023448e1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd17172fd_996e_496e_92b5_9cd19edc65f4.slice/crio-b341537ccb4b136de4ee6b2f73d91bb94b0972196b2cc9404f3b92ace4936702\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d518ac2_8f5c_4520_b3b4_8e01f6678974.slice/crio-1f135306da0aaa11ef00c9f9f2c78a4ba01695e533b490b69523813a529507ce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod718f5a48_c119_4bae_9c9c_72ed544758ee.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86426b6e_ec55_49d4_804e_75b204239fa1.slice/crio-d8b69c7107b0c5f91af953d758844dc3ee89532cd4ca4ee00b6b4829ad5f3f18.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d518ac2_8f5c_4520_b3b4_8e01f6678974.slice/crio-6f3a2f449fb65e19cac86104789c3ad946517e7dd0bf835c4455fa34962b7df9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd17172fd_996e_496e_92b5_9cd19edc65f4.slice/crio-94848df684c7d16e35f3425fecf44bad6bb45f6cba749b60673a0a0b7fe71222.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd680dc75_619e_41ae_90cb_2f58898a33ac.slice/crio-0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22f8c42f_134c_47ed_baa8_0f176a362d38.slice/crio-conmon-fb68641df086d68c697275b20dc2997ea82074be958cf89d4c1031793e6d4fcb.scope\": RecentStats: unable to find data in memory cache]" Mar 17 16:04:10 crc kubenswrapper[4767]: E0317 16:04:10.767002 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff0f3ad0_1b55_4abb_9eb6_c07d82251bdd.slice/crio-2055be26e06e56bc2b2a7957901aafdce92a48fee3057028796f99ee80231978.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86426b6e_ec55_49d4_804e_75b204239fa1.slice/crio-conmon-d8b69c7107b0c5f91af953d758844dc3ee89532cd4ca4ee00b6b4829ad5f3f18.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86426b6e_ec55_49d4_804e_75b204239fa1.slice/crio-c2b9c871dd75268d8313f0a8947b962e08fc00dd9b8852847187eaeebe41cf7d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d518ac2_8f5c_4520_b3b4_8e01f6678974.slice/crio-conmon-6f3a2f449fb65e19cac86104789c3ad946517e7dd0bf835c4455fa34962b7df9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod718f5a48_c119_4bae_9c9c_72ed544758ee.slice/crio-conmon-7cbcbb684b7117829363a6d6637dc12889f3744a6903dce8fbfb35e6f52bd8f8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4676bc_09a3_4b2d_9bc9_579b497bbfaf.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod718f5a48_c119_4bae_9c9c_72ed544758ee.slice/crio-7cbcbb684b7117829363a6d6637dc12889f3744a6903dce8fbfb35e6f52bd8f8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod718f5a48_c119_4bae_9c9c_72ed544758ee.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d518ac2_8f5c_4520_b3b4_8e01f6678974.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d518ac2_8f5c_4520_b3b4_8e01f6678974.slice/crio-1f135306da0aaa11ef00c9f9f2c78a4ba01695e533b490b69523813a529507ce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d518ac2_8f5c_4520_b3b4_8e01f6678974.slice/crio-conmon-1f135306da0aaa11ef00c9f9f2c78a4ba01695e533b490b69523813a529507ce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff0f3ad0_1b55_4abb_9eb6_c07d82251bdd.slice/crio-conmon-2055be26e06e56bc2b2a7957901aafdce92a48fee3057028796f99ee80231978.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc95699d_4853_4d4b_9239_019601fd7df8.slice/crio-1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc95699d_4853_4d4b_9239_019601fd7df8.slice/crio-conmon-1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86426b6e_ec55_49d4_804e_75b204239fa1.slice/crio-conmon-c2b9c871dd75268d8313f0a8947b962e08fc00dd9b8852847187eaeebe41cf7d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod718f5a48_c119_4bae_9c9c_72ed544758ee.slice/crio-54dec61b8f1f3cbb7d95d23d5348436aae4cab3a396a7a3433084e138d2d84f2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d518ac2_8f5c_4520_b3b4_8e01f6678974.slice/crio-6f3a2f449fb65e19cac86104789c3ad946517e7dd0bf835c4455fa34962b7df9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd680dc75_619e_41ae_90cb_2f58898a33ac.slice/crio-conmon-0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86426b6e_ec55_49d4_804e_75b204239fa1.slice/crio-d8b69c7107b0c5f91af953d758844dc3ee89532cd4ca4ee00b6b4829ad5f3f18.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd680dc75_619e_41ae_90cb_2f58898a33ac.slice/crio-0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c.scope\": RecentStats: unable to find data in memory cache]" Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.882665 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-ovndb-tls-certs\") pod \"86426b6e-ec55-49d4-804e-75b204239fa1\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.882841 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rlgp\" (UniqueName: \"kubernetes.io/projected/86426b6e-ec55-49d4-804e-75b204239fa1-kube-api-access-8rlgp\") pod \"86426b6e-ec55-49d4-804e-75b204239fa1\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.882896 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-combined-ca-bundle\") pod \"86426b6e-ec55-49d4-804e-75b204239fa1\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.883075 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-config\") pod \"86426b6e-ec55-49d4-804e-75b204239fa1\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.883261 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-httpd-config\") pod \"86426b6e-ec55-49d4-804e-75b204239fa1\" (UID: \"86426b6e-ec55-49d4-804e-75b204239fa1\") " Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.891692 4767 scope.go:117] "RemoveContainer" containerID="66f633e8bd912a5ee681ce4f080954b74553019edf6c5e56558a68eed3012220" Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.922838 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "86426b6e-ec55-49d4-804e-75b204239fa1" (UID: "86426b6e-ec55-49d4-804e-75b204239fa1"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.924105 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86426b6e-ec55-49d4-804e-75b204239fa1-kube-api-access-8rlgp" (OuterVolumeSpecName: "kube-api-access-8rlgp") pod "86426b6e-ec55-49d4-804e-75b204239fa1" (UID: "86426b6e-ec55-49d4-804e-75b204239fa1"). InnerVolumeSpecName "kube-api-access-8rlgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.995698 4767 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:10 crc kubenswrapper[4767]: I0317 16:04:10.996025 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rlgp\" (UniqueName: \"kubernetes.io/projected/86426b6e-ec55-49d4-804e-75b204239fa1-kube-api-access-8rlgp\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.002299 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86426b6e-ec55-49d4-804e-75b204239fa1" (UID: "86426b6e-ec55-49d4-804e-75b204239fa1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.008088 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.098632 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data-custom\") pod \"dc95699d-4853-4d4b-9239-019601fd7df8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.098780 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn29m\" (UniqueName: \"kubernetes.io/projected/dc95699d-4853-4d4b-9239-019601fd7df8-kube-api-access-xn29m\") pod \"dc95699d-4853-4d4b-9239-019601fd7df8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.098929 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc95699d-4853-4d4b-9239-019601fd7df8-logs\") pod \"dc95699d-4853-4d4b-9239-019601fd7df8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.098956 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-combined-ca-bundle\") pod \"dc95699d-4853-4d4b-9239-019601fd7df8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.099348 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data\") pod \"dc95699d-4853-4d4b-9239-019601fd7df8\" (UID: \"dc95699d-4853-4d4b-9239-019601fd7df8\") " Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.099714 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc95699d-4853-4d4b-9239-019601fd7df8-logs" (OuterVolumeSpecName: "logs") pod "dc95699d-4853-4d4b-9239-019601fd7df8" (UID: "dc95699d-4853-4d4b-9239-019601fd7df8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.111855 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc95699d-4853-4d4b-9239-019601fd7df8-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.111906 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.115639 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "dc95699d-4853-4d4b-9239-019601fd7df8" (UID: "dc95699d-4853-4d4b-9239-019601fd7df8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.115674 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc95699d-4853-4d4b-9239-019601fd7df8-kube-api-access-xn29m" (OuterVolumeSpecName: "kube-api-access-xn29m") pod "dc95699d-4853-4d4b-9239-019601fd7df8" (UID: "dc95699d-4853-4d4b-9239-019601fd7df8"). InnerVolumeSpecName "kube-api-access-xn29m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.224922 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.224974 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn29m\" (UniqueName: \"kubernetes.io/projected/dc95699d-4853-4d4b-9239-019601fd7df8-kube-api-access-xn29m\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.225042 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-config" (OuterVolumeSpecName: "config") pod "86426b6e-ec55-49d4-804e-75b204239fa1" (UID: "86426b6e-ec55-49d4-804e-75b204239fa1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.226492 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc95699d-4853-4d4b-9239-019601fd7df8" (UID: "dc95699d-4853-4d4b-9239-019601fd7df8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.232210 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "86426b6e-ec55-49d4-804e-75b204239fa1" (UID: "86426b6e-ec55-49d4-804e-75b204239fa1"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.257965 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-667f69cbdb-5qvs7" event={"ID":"86426b6e-ec55-49d4-804e-75b204239fa1","Type":"ContainerDied","Data":"12eeed4dfedd2c540c520f5721a3b510295e1093b21b2b1ca3a1610081bf9ad0"} Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.258362 4767 scope.go:117] "RemoveContainer" containerID="d8b69c7107b0c5f91af953d758844dc3ee89532cd4ca4ee00b6b4829ad5f3f18" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.258600 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-667f69cbdb-5qvs7" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.293780 4767 generic.go:334] "Generic (PLEG): container finished" podID="ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" containerID="2055be26e06e56bc2b2a7957901aafdce92a48fee3057028796f99ee80231978" exitCode=137 Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.293858 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" event={"ID":"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd","Type":"ContainerDied","Data":"2055be26e06e56bc2b2a7957901aafdce92a48fee3057028796f99ee80231978"} Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.320371 4767 generic.go:334] "Generic (PLEG): container finished" podID="dc95699d-4853-4d4b-9239-019601fd7df8" containerID="1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325" exitCode=137 Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.320475 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f6f54698c-rrhd8" event={"ID":"dc95699d-4853-4d4b-9239-019601fd7df8","Type":"ContainerDied","Data":"1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325"} Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.320507 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f6f54698c-rrhd8" event={"ID":"dc95699d-4853-4d4b-9239-019601fd7df8","Type":"ContainerDied","Data":"ef67e9fd96b6864bdce50fffd18441a5cc8a6f39f638ebcc13ab497c753751cb"} Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.320619 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7f6f54698c-rrhd8" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.327179 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.327416 4767 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/86426b6e-ec55-49d4-804e-75b204239fa1-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.327526 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.338348 4767 scope.go:117] "RemoveContainer" containerID="c2b9c871dd75268d8313f0a8947b962e08fc00dd9b8852847187eaeebe41cf7d" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.349180 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data" (OuterVolumeSpecName: "config-data") pod "dc95699d-4853-4d4b-9239-019601fd7df8" (UID: "dc95699d-4853-4d4b-9239-019601fd7df8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.382834 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25debbd1-41e4-42db-af74-c058c0d6d9b2" path="/var/lib/kubelet/pods/25debbd1-41e4-42db-af74-c058c0d6d9b2/volumes" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.386391 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a248ba4a-e1d4-4204-b549-7dfff52fe100" path="/var/lib/kubelet/pods/a248ba4a-e1d4-4204-b549-7dfff52fe100/volumes" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.424098 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.428342 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-667f69cbdb-5qvs7"] Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.431555 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc95699d-4853-4d4b-9239-019601fd7df8-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.446998 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-667f69cbdb-5qvs7"] Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.449625 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.493132 4767 scope.go:117] "RemoveContainer" containerID="1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.534353 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgr7z\" (UniqueName: \"kubernetes.io/projected/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-kube-api-access-tgr7z\") pod \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.534498 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-logs\") pod \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.534721 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data-custom\") pod \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.534873 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-combined-ca-bundle\") pod \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.534926 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data\") pod \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\" (UID: \"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd\") " Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.535755 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-logs" (OuterVolumeSpecName: "logs") pod "ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" (UID: "ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.541015 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-kube-api-access-tgr7z" (OuterVolumeSpecName: "kube-api-access-tgr7z") pod "ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" (UID: "ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd"). InnerVolumeSpecName "kube-api-access-tgr7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.565857 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" (UID: "ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.585902 4767 scope.go:117] "RemoveContainer" containerID="295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.616592 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" (UID: "ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.637870 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.639117 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.639217 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgr7z\" (UniqueName: \"kubernetes.io/projected/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-kube-api-access-tgr7z\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.639288 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.666626 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7f6f54698c-rrhd8"] Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.693592 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data" (OuterVolumeSpecName: "config-data") pod "ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" (UID: "ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.704452 4767 scope.go:117] "RemoveContainer" containerID="1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325" Mar 17 16:04:11 crc kubenswrapper[4767]: E0317 16:04:11.705566 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325\": container with ID starting with 1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325 not found: ID does not exist" containerID="1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.706477 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325"} err="failed to get container status \"1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325\": rpc error: code = NotFound desc = could not find container \"1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325\": container with ID starting with 1e380664b9b766e91058615dbce070866363ee44b7d9dcab8cd08beb30471325 not found: ID does not exist" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.706591 4767 scope.go:117] "RemoveContainer" containerID="295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c" Mar 17 16:04:11 crc kubenswrapper[4767]: E0317 16:04:11.708658 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c\": container with ID starting with 295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c not found: ID does not exist" containerID="295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.708739 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c"} err="failed to get container status \"295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c\": rpc error: code = NotFound desc = could not find container \"295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c\": container with ID starting with 295029fd66d4df71c76847710035db864c165c6248723828445ae1bcf310c92c not found: ID does not exist" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.755561 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:11 crc kubenswrapper[4767]: I0317 16:04:11.758105 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-7f6f54698c-rrhd8"] Mar 17 16:04:12 crc kubenswrapper[4767]: I0317 16:04:12.421753 4767 generic.go:334] "Generic (PLEG): container finished" podID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerID="d578683d641746e5a92c34391fd4826e7a2330cdbaf5dabe1c682894cf1d52c6" exitCode=0 Mar 17 16:04:12 crc kubenswrapper[4767]: I0317 16:04:12.421868 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crqkh" event={"ID":"4ad1c7b7-377f-488c-af38-bf30a79c3589","Type":"ContainerDied","Data":"d578683d641746e5a92c34391fd4826e7a2330cdbaf5dabe1c682894cf1d52c6"} Mar 17 16:04:12 crc kubenswrapper[4767]: I0317 16:04:12.512592 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" event={"ID":"ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd","Type":"ContainerDied","Data":"6a7698f4045ba558680b51081489fc574b78c028bcaebe433f4425c74b3e6b03"} Mar 17 16:04:12 crc kubenswrapper[4767]: I0317 16:04:12.512665 4767 scope.go:117] "RemoveContainer" containerID="2055be26e06e56bc2b2a7957901aafdce92a48fee3057028796f99ee80231978" Mar 17 16:04:12 crc kubenswrapper[4767]: I0317 16:04:12.512872 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-579bdf4fdb-kq7rd" Mar 17 16:04:12 crc kubenswrapper[4767]: I0317 16:04:12.646099 4767 scope.go:117] "RemoveContainer" containerID="eafb36bee52ba0fdd7580179e1afc8a82c42db16efb072a002849fddd3ae245d" Mar 17 16:04:12 crc kubenswrapper[4767]: I0317 16:04:12.674915 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-579bdf4fdb-kq7rd"] Mar 17 16:04:12 crc kubenswrapper[4767]: I0317 16:04:12.691288 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-579bdf4fdb-kq7rd"] Mar 17 16:04:13 crc kubenswrapper[4767]: I0317 16:04:13.386149 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86426b6e-ec55-49d4-804e-75b204239fa1" path="/var/lib/kubelet/pods/86426b6e-ec55-49d4-804e-75b204239fa1/volumes" Mar 17 16:04:13 crc kubenswrapper[4767]: I0317 16:04:13.387352 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc95699d-4853-4d4b-9239-019601fd7df8" path="/var/lib/kubelet/pods/dc95699d-4853-4d4b-9239-019601fd7df8/volumes" Mar 17 16:04:13 crc kubenswrapper[4767]: I0317 16:04:13.388306 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" path="/var/lib/kubelet/pods/ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd/volumes" Mar 17 16:04:13 crc kubenswrapper[4767]: I0317 16:04:13.668909 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crqkh" event={"ID":"4ad1c7b7-377f-488c-af38-bf30a79c3589","Type":"ContainerStarted","Data":"1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770"} Mar 17 16:04:13 crc kubenswrapper[4767]: E0317 16:04:13.692071 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:04:13 crc kubenswrapper[4767]: I0317 16:04:13.699777 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-crqkh" podStartSLOduration=3.7685754559999998 podStartE2EDuration="6.699746192s" podCreationTimestamp="2026-03-17 16:04:07 +0000 UTC" firstStartedPulling="2026-03-17 16:04:09.984765917 +0000 UTC m=+1641.398081954" lastFinishedPulling="2026-03-17 16:04:12.915936643 +0000 UTC m=+1644.329252690" observedRunningTime="2026-03-17 16:04:13.69268663 +0000 UTC m=+1645.106002697" watchObservedRunningTime="2026-03-17 16:04:13.699746192 +0000 UTC m=+1645.113062239" Mar 17 16:04:14 crc kubenswrapper[4767]: I0317 16:04:14.235434 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="2ca1d252-faff-4972-8ca6-5b66cf11de53" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.216:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 16:04:16 crc kubenswrapper[4767]: I0317 16:04:16.901639 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 17 16:04:17 crc kubenswrapper[4767]: I0317 16:04:17.710732 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:17 crc kubenswrapper[4767]: I0317 16:04:17.711079 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:17 crc kubenswrapper[4767]: I0317 16:04:17.806318 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:17 crc kubenswrapper[4767]: I0317 16:04:17.824307 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:04:17 crc kubenswrapper[4767]: I0317 16:04:17.824753 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="ceilometer-central-agent" containerID="cri-o://b673dbe3c730b2a2acdf223e799163fe36598b87eb20e0ad865592a2c81c71ac" gracePeriod=30 Mar 17 16:04:17 crc kubenswrapper[4767]: I0317 16:04:17.824952 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="ceilometer-notification-agent" containerID="cri-o://dfe0080987b662004c5cc0a001db82ea1328fb2304d1b10898150fe5a5f55213" gracePeriod=30 Mar 17 16:04:17 crc kubenswrapper[4767]: I0317 16:04:17.825010 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="sg-core" containerID="cri-o://e2a40d99e8f68af20b6fa59961e265d0c3832a7052b6c49304e606de6ba4d2da" gracePeriod=30 Mar 17 16:04:17 crc kubenswrapper[4767]: I0317 16:04:17.825044 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="proxy-httpd" containerID="cri-o://b4763a53b116968d26349a0361954feba8bb8491eb1e171954857b214c937eb0" gracePeriod=30 Mar 17 16:04:18 crc kubenswrapper[4767]: I0317 16:04:18.462667 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 17 16:04:18 crc kubenswrapper[4767]: I0317 16:04:18.804901 4767 generic.go:334] "Generic (PLEG): container finished" podID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerID="b4763a53b116968d26349a0361954feba8bb8491eb1e171954857b214c937eb0" exitCode=0 Mar 17 16:04:18 crc kubenswrapper[4767]: I0317 16:04:18.807708 4767 generic.go:334] "Generic (PLEG): container finished" podID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerID="e2a40d99e8f68af20b6fa59961e265d0c3832a7052b6c49304e606de6ba4d2da" exitCode=2 Mar 17 16:04:18 crc kubenswrapper[4767]: I0317 16:04:18.807729 4767 generic.go:334] "Generic (PLEG): container finished" podID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerID="dfe0080987b662004c5cc0a001db82ea1328fb2304d1b10898150fe5a5f55213" exitCode=0 Mar 17 16:04:18 crc kubenswrapper[4767]: I0317 16:04:18.807736 4767 generic.go:334] "Generic (PLEG): container finished" podID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerID="b673dbe3c730b2a2acdf223e799163fe36598b87eb20e0ad865592a2c81c71ac" exitCode=0 Mar 17 16:04:18 crc kubenswrapper[4767]: I0317 16:04:18.805317 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d2da299-4e48-4135-a749-7a2adf3f7ec9","Type":"ContainerDied","Data":"b4763a53b116968d26349a0361954feba8bb8491eb1e171954857b214c937eb0"} Mar 17 16:04:18 crc kubenswrapper[4767]: I0317 16:04:18.809283 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d2da299-4e48-4135-a749-7a2adf3f7ec9","Type":"ContainerDied","Data":"e2a40d99e8f68af20b6fa59961e265d0c3832a7052b6c49304e606de6ba4d2da"} Mar 17 16:04:18 crc kubenswrapper[4767]: I0317 16:04:18.809313 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d2da299-4e48-4135-a749-7a2adf3f7ec9","Type":"ContainerDied","Data":"dfe0080987b662004c5cc0a001db82ea1328fb2304d1b10898150fe5a5f55213"} Mar 17 16:04:18 crc kubenswrapper[4767]: I0317 16:04:18.809330 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d2da299-4e48-4135-a749-7a2adf3f7ec9","Type":"ContainerDied","Data":"b673dbe3c730b2a2acdf223e799163fe36598b87eb20e0ad865592a2c81c71ac"} Mar 17 16:04:18 crc kubenswrapper[4767]: I0317 16:04:18.883363 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.026011 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-crqkh"] Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.379627 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:04:19 crc kubenswrapper[4767]: E0317 16:04:19.381527 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.493149 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-66ff489c4c-47nsf"] Mar 17 16:04:19 crc kubenswrapper[4767]: E0317 16:04:19.494102 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc95699d-4853-4d4b-9239-019601fd7df8" containerName="barbican-worker" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494119 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc95699d-4853-4d4b-9239-019601fd7df8" containerName="barbican-worker" Mar 17 16:04:19 crc kubenswrapper[4767]: E0317 16:04:19.494142 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25debbd1-41e4-42db-af74-c058c0d6d9b2" containerName="registry-server" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494148 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="25debbd1-41e4-42db-af74-c058c0d6d9b2" containerName="registry-server" Mar 17 16:04:19 crc kubenswrapper[4767]: E0317 16:04:19.494159 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc95699d-4853-4d4b-9239-019601fd7df8" containerName="barbican-worker-log" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494169 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc95699d-4853-4d4b-9239-019601fd7df8" containerName="barbican-worker-log" Mar 17 16:04:19 crc kubenswrapper[4767]: E0317 16:04:19.494206 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25debbd1-41e4-42db-af74-c058c0d6d9b2" containerName="extract-utilities" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494212 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="25debbd1-41e4-42db-af74-c058c0d6d9b2" containerName="extract-utilities" Mar 17 16:04:19 crc kubenswrapper[4767]: E0317 16:04:19.494239 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86426b6e-ec55-49d4-804e-75b204239fa1" containerName="neutron-httpd" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494245 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="86426b6e-ec55-49d4-804e-75b204239fa1" containerName="neutron-httpd" Mar 17 16:04:19 crc kubenswrapper[4767]: E0317 16:04:19.494258 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25debbd1-41e4-42db-af74-c058c0d6d9b2" containerName="extract-content" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494264 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="25debbd1-41e4-42db-af74-c058c0d6d9b2" containerName="extract-content" Mar 17 16:04:19 crc kubenswrapper[4767]: E0317 16:04:19.494277 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b1f5965-725f-47c1-bfc3-cdb8a3b7a696" containerName="oc" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494283 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b1f5965-725f-47c1-bfc3-cdb8a3b7a696" containerName="oc" Mar 17 16:04:19 crc kubenswrapper[4767]: E0317 16:04:19.494300 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86426b6e-ec55-49d4-804e-75b204239fa1" containerName="neutron-api" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494305 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="86426b6e-ec55-49d4-804e-75b204239fa1" containerName="neutron-api" Mar 17 16:04:19 crc kubenswrapper[4767]: E0317 16:04:19.494316 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" containerName="barbican-keystone-listener" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494322 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" containerName="barbican-keystone-listener" Mar 17 16:04:19 crc kubenswrapper[4767]: E0317 16:04:19.494335 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" containerName="barbican-keystone-listener-log" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494341 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" containerName="barbican-keystone-listener-log" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494575 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b1f5965-725f-47c1-bfc3-cdb8a3b7a696" containerName="oc" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494587 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="86426b6e-ec55-49d4-804e-75b204239fa1" containerName="neutron-httpd" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494605 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="25debbd1-41e4-42db-af74-c058c0d6d9b2" containerName="registry-server" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494623 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" containerName="barbican-keystone-listener-log" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494639 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc95699d-4853-4d4b-9239-019601fd7df8" containerName="barbican-worker" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494647 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="86426b6e-ec55-49d4-804e-75b204239fa1" containerName="neutron-api" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494657 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc95699d-4853-4d4b-9239-019601fd7df8" containerName="barbican-worker-log" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.494674 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff0f3ad0-1b55-4abb-9eb6-c07d82251bdd" containerName="barbican-keystone-listener" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.499747 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.507234 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.507583 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-kmsxr" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.507746 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.555412 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-66ff489c4c-47nsf"] Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.618814 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gshrj\" (UniqueName: \"kubernetes.io/projected/2610e145-1410-4330-bcca-808faf035214-kube-api-access-gshrj\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.618917 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-combined-ca-bundle\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.619032 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data-custom\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.619065 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.635920 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-q896r"] Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.638609 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.651863 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-q896r"] Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.721090 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbxsm\" (UniqueName: \"kubernetes.io/projected/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-kube-api-access-sbxsm\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.721361 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-combined-ca-bundle\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.721446 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.721511 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-config\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.721572 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.721634 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data-custom\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.721678 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.721779 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.721812 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.721838 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gshrj\" (UniqueName: \"kubernetes.io/projected/2610e145-1410-4330-bcca-808faf035214-kube-api-access-gshrj\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.737763 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.747306 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data-custom\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.758450 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-combined-ca-bundle\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.768446 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gshrj\" (UniqueName: \"kubernetes.io/projected/2610e145-1410-4330-bcca-808faf035214-kube-api-access-gshrj\") pod \"heat-engine-66ff489c4c-47nsf\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.803299 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-658d967468-8hprg"] Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.805923 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.820483 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.824394 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.845282 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-config\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.845399 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.845745 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data-custom\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.845806 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.835958 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-658d967468-8hprg"] Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.828322 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.842015 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.865149 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.865342 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-combined-ca-bundle\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.865407 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbxsm\" (UniqueName: \"kubernetes.io/projected/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-kube-api-access-sbxsm\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.865448 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9twc\" (UniqueName: \"kubernetes.io/projected/00a982e7-017e-4f3c-a8aa-21eab359cad4-kube-api-access-r9twc\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.865564 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.865915 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-config\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.866790 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.870809 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-75777db76d-5z4qz"] Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.871377 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.877420 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.885960 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.897785 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.940877 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbxsm\" (UniqueName: \"kubernetes.io/projected/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-kube-api-access-sbxsm\") pod \"dnsmasq-dns-7756b9d78c-q896r\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.969619 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-75777db76d-5z4qz"] Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.990312 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.990667 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data-custom\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.990747 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-combined-ca-bundle\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.990817 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data-custom\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.990923 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-combined-ca-bundle\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.990986 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9twc\" (UniqueName: \"kubernetes.io/projected/00a982e7-017e-4f3c-a8aa-21eab359cad4-kube-api-access-r9twc\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.991094 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:19 crc kubenswrapper[4767]: I0317 16:04:19.991155 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tfkg\" (UniqueName: \"kubernetes.io/projected/3caf16d7-044a-4efc-a3ba-c494255df8ff-kube-api-access-7tfkg\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.000510 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-combined-ca-bundle\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.005472 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.013468 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data-custom\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.013704 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9twc\" (UniqueName: \"kubernetes.io/projected/00a982e7-017e-4f3c-a8aa-21eab359cad4-kube-api-access-r9twc\") pod \"heat-cfnapi-658d967468-8hprg\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.014769 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.094153 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-combined-ca-bundle\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.094234 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data-custom\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.094307 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tfkg\" (UniqueName: \"kubernetes.io/projected/3caf16d7-044a-4efc-a3ba-c494255df8ff-kube-api-access-7tfkg\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.094416 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.102318 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data-custom\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.105632 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-combined-ca-bundle\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.113751 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.127367 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tfkg\" (UniqueName: \"kubernetes.io/projected/3caf16d7-044a-4efc-a3ba-c494255df8ff-kube-api-access-7tfkg\") pod \"heat-api-75777db76d-5z4qz\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.200304 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.246203 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.594476 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-56879488c7-5jbhc"] Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.597777 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.603692 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.608096 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-56879488c7-5jbhc"] Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.611355 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.630073 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.731837 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65779f05-d74b-49a7-a31a-2de7f9005e59-log-httpd\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.732464 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/65779f05-d74b-49a7-a31a-2de7f9005e59-etc-swift\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.732497 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-internal-tls-certs\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.734705 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-combined-ca-bundle\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.734805 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-public-tls-certs\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.734871 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-config-data\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.734976 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp5fd\" (UniqueName: \"kubernetes.io/projected/65779f05-d74b-49a7-a31a-2de7f9005e59-kube-api-access-zp5fd\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.735304 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65779f05-d74b-49a7-a31a-2de7f9005e59-run-httpd\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.838846 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp5fd\" (UniqueName: \"kubernetes.io/projected/65779f05-d74b-49a7-a31a-2de7f9005e59-kube-api-access-zp5fd\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.839117 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65779f05-d74b-49a7-a31a-2de7f9005e59-run-httpd\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.839454 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65779f05-d74b-49a7-a31a-2de7f9005e59-log-httpd\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.839481 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/65779f05-d74b-49a7-a31a-2de7f9005e59-etc-swift\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.839513 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-internal-tls-certs\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.839793 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-combined-ca-bundle\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.839843 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-public-tls-certs\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.840429 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65779f05-d74b-49a7-a31a-2de7f9005e59-run-httpd\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.840480 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65779f05-d74b-49a7-a31a-2de7f9005e59-log-httpd\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.839889 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-config-data\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.846322 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-public-tls-certs\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.849051 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/65779f05-d74b-49a7-a31a-2de7f9005e59-etc-swift\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.850169 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-config-data\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.850796 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-internal-tls-certs\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.864074 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp5fd\" (UniqueName: \"kubernetes.io/projected/65779f05-d74b-49a7-a31a-2de7f9005e59-kube-api-access-zp5fd\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.865109 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65779f05-d74b-49a7-a31a-2de7f9005e59-combined-ca-bundle\") pod \"swift-proxy-56879488c7-5jbhc\" (UID: \"65779f05-d74b-49a7-a31a-2de7f9005e59\") " pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.882579 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-crqkh" podUID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerName="registry-server" containerID="cri-o://1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770" gracePeriod=2 Mar 17 16:04:20 crc kubenswrapper[4767]: I0317 16:04:20.930416 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:21 crc kubenswrapper[4767]: E0317 16:04:21.229636 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ad1c7b7_377f_488c_af38_bf30a79c3589.slice/crio-conmon-1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ad1c7b7_377f_488c_af38_bf30a79c3589.slice/crio-1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770.scope\": RecentStats: unable to find data in memory cache]" Mar 17 16:04:21 crc kubenswrapper[4767]: I0317 16:04:21.933364 4767 generic.go:334] "Generic (PLEG): container finished" podID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerID="1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770" exitCode=0 Mar 17 16:04:21 crc kubenswrapper[4767]: I0317 16:04:21.933494 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crqkh" event={"ID":"4ad1c7b7-377f-488c-af38-bf30a79c3589","Type":"ContainerDied","Data":"1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770"} Mar 17 16:04:27 crc kubenswrapper[4767]: E0317 16:04:27.742069 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770 is running failed: container process not found" containerID="1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 16:04:27 crc kubenswrapper[4767]: E0317 16:04:27.743152 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770 is running failed: container process not found" containerID="1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 16:04:27 crc kubenswrapper[4767]: E0317 16:04:27.745477 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770 is running failed: container process not found" containerID="1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 16:04:27 crc kubenswrapper[4767]: E0317 16:04:27.745578 4767 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-crqkh" podUID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerName="registry-server" Mar 17 16:04:28 crc kubenswrapper[4767]: E0317 16:04:28.982603 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.492205 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5bdf466bd4-x9rf5"] Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.694580 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5bdf466bd4-x9rf5"] Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.694730 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.796409 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-58b6b75dc-pjv6d"] Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.801744 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.806103 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86lfx\" (UniqueName: \"kubernetes.io/projected/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-kube-api-access-86lfx\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.806905 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data-custom\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.807048 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-combined-ca-bundle\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.807408 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.831114 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-9559455fd-vntnw"] Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.833103 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.855489 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-9559455fd-vntnw"] Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.881523 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58b6b75dc-pjv6d"] Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.910573 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.910664 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-combined-ca-bundle\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.910724 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data-custom\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.910783 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9rfw\" (UniqueName: \"kubernetes.io/projected/33f239ab-85ed-4803-a0db-c31dc5301d28-kube-api-access-t9rfw\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.910823 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86lfx\" (UniqueName: \"kubernetes.io/projected/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-kube-api-access-86lfx\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.910935 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data-custom\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.910976 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data-custom\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.911007 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-combined-ca-bundle\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.911155 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95z8k\" (UniqueName: \"kubernetes.io/projected/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-kube-api-access-95z8k\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.911295 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-combined-ca-bundle\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.911344 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.911404 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.920245 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-combined-ca-bundle\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.920725 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.936556 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data-custom\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:29 crc kubenswrapper[4767]: I0317 16:04:29.940203 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86lfx\" (UniqueName: \"kubernetes.io/projected/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-kube-api-access-86lfx\") pod \"heat-engine-5bdf466bd4-x9rf5\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.013617 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.013706 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.013755 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-combined-ca-bundle\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.013802 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data-custom\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.013849 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9rfw\" (UniqueName: \"kubernetes.io/projected/33f239ab-85ed-4803-a0db-c31dc5301d28-kube-api-access-t9rfw\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.013942 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data-custom\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.014042 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95z8k\" (UniqueName: \"kubernetes.io/projected/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-kube-api-access-95z8k\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.014085 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-combined-ca-bundle\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.019811 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data-custom\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.019926 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data-custom\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.027577 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-combined-ca-bundle\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.028504 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-combined-ca-bundle\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.029739 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.031849 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.058954 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95z8k\" (UniqueName: \"kubernetes.io/projected/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-kube-api-access-95z8k\") pod \"heat-cfnapi-58b6b75dc-pjv6d\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.084068 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9rfw\" (UniqueName: \"kubernetes.io/projected/33f239ab-85ed-4803-a0db-c31dc5301d28-kube-api-access-t9rfw\") pod \"heat-api-9559455fd-vntnw\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.112067 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.135020 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.166079 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:30 crc kubenswrapper[4767]: I0317 16:04:30.375872 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:04:30 crc kubenswrapper[4767]: E0317 16:04:30.379020 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:04:31 crc kubenswrapper[4767]: E0317 16:04:31.300831 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.359757 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-75777db76d-5z4qz"] Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.401869 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6449cd4787-bslls"] Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.403915 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.406806 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.407479 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.440898 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-658d967468-8hprg"] Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.464326 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data-custom\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.464469 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-internal-tls-certs\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.464526 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-combined-ca-bundle\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.464664 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.464705 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-public-tls-certs\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.465574 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjqt8\" (UniqueName: \"kubernetes.io/projected/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-kube-api-access-gjqt8\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.485653 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6449cd4787-bslls"] Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.709128 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjqt8\" (UniqueName: \"kubernetes.io/projected/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-kube-api-access-gjqt8\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.709213 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data-custom\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.709301 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-internal-tls-certs\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.709367 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-combined-ca-bundle\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.709530 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.709569 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-public-tls-certs\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.721757 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-combined-ca-bundle\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.722711 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data-custom\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.723956 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.727794 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-internal-tls-certs\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.729628 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-public-tls-certs\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.737262 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-56f748dcbf-9x7mb"] Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.739697 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.748249 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-56f748dcbf-9x7mb"] Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.750109 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.750450 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.778901 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjqt8\" (UniqueName: \"kubernetes.io/projected/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-kube-api-access-gjqt8\") pod \"heat-api-6449cd4787-bslls\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.812255 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-combined-ca-bundle\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.813196 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-public-tls-certs\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.813388 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.814386 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-internal-tls-certs\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.814555 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data-custom\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.814712 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gt24\" (UniqueName: \"kubernetes.io/projected/43b3b246-f5d3-4311-ab67-d30e53ac7082-kube-api-access-2gt24\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: E0317 16:04:32.845806 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Mar 17 16:04:32 crc kubenswrapper[4767]: E0317 16:04:32.846341 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n54bh5cch56bh675h594h5c6h646h546h594hb8h678h57dh68ch65fh85h566hb8hfchbdh66bhcdh56chf5h546h69h64fhfh5cdhf6h4h658h84q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_CA_CERT,Value:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dbldk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(049b9f3b-c6db-4274-9f45-c5cac0c0a17b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:04:32 crc kubenswrapper[4767]: E0317 16:04:32.848235 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="049b9f3b-c6db-4274-9f45-c5cac0c0a17b" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.916832 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-internal-tls-certs\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.916888 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data-custom\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.916954 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gt24\" (UniqueName: \"kubernetes.io/projected/43b3b246-f5d3-4311-ab67-d30e53ac7082-kube-api-access-2gt24\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.917063 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-combined-ca-bundle\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.917212 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-public-tls-certs\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.921947 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.925611 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data-custom\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.929579 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-combined-ca-bundle\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.929638 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-public-tls-certs\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.939913 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-internal-tls-certs\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.941288 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:32 crc kubenswrapper[4767]: I0317 16:04:32.950774 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gt24\" (UniqueName: \"kubernetes.io/projected/43b3b246-f5d3-4311-ab67-d30e53ac7082-kube-api-access-2gt24\") pod \"heat-cfnapi-56f748dcbf-9x7mb\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.070710 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.089489 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.134771 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.156234 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.156337 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d2da299-4e48-4135-a749-7a2adf3f7ec9","Type":"ContainerDied","Data":"fbdf5f458101f27459fc0ac59386d618852caa139b35a854248efa096544ebfa"} Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.156429 4767 scope.go:117] "RemoveContainer" containerID="b4763a53b116968d26349a0361954feba8bb8491eb1e171954857b214c937eb0" Mar 17 16:04:33 crc kubenswrapper[4767]: E0317 16:04:33.162832 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="049b9f3b-c6db-4274-9f45-c5cac0c0a17b" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.247799 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-config-data\") pod \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.248303 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-log-httpd\") pod \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.248351 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-scripts\") pod \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.248485 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-combined-ca-bundle\") pod \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.248518 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-sg-core-conf-yaml\") pod \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.248631 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gzps\" (UniqueName: \"kubernetes.io/projected/0d2da299-4e48-4135-a749-7a2adf3f7ec9-kube-api-access-5gzps\") pod \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.248767 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-run-httpd\") pod \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\" (UID: \"0d2da299-4e48-4135-a749-7a2adf3f7ec9\") " Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.260712 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0d2da299-4e48-4135-a749-7a2adf3f7ec9" (UID: "0d2da299-4e48-4135-a749-7a2adf3f7ec9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.274121 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0d2da299-4e48-4135-a749-7a2adf3f7ec9" (UID: "0d2da299-4e48-4135-a749-7a2adf3f7ec9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.323226 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-scripts" (OuterVolumeSpecName: "scripts") pod "0d2da299-4e48-4135-a749-7a2adf3f7ec9" (UID: "0d2da299-4e48-4135-a749-7a2adf3f7ec9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.323570 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d2da299-4e48-4135-a749-7a2adf3f7ec9-kube-api-access-5gzps" (OuterVolumeSpecName: "kube-api-access-5gzps") pod "0d2da299-4e48-4135-a749-7a2adf3f7ec9" (UID: "0d2da299-4e48-4135-a749-7a2adf3f7ec9"). InnerVolumeSpecName "kube-api-access-5gzps". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.550231 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0d2da299-4e48-4135-a749-7a2adf3f7ec9" (UID: "0d2da299-4e48-4135-a749-7a2adf3f7ec9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.594873 4767 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.595885 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.596121 4767 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.596210 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gzps\" (UniqueName: \"kubernetes.io/projected/0d2da299-4e48-4135-a749-7a2adf3f7ec9-kube-api-access-5gzps\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.596223 4767 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d2da299-4e48-4135-a749-7a2adf3f7ec9-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.857433 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d2da299-4e48-4135-a749-7a2adf3f7ec9" (UID: "0d2da299-4e48-4135-a749-7a2adf3f7ec9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.907487 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-config-data" (OuterVolumeSpecName: "config-data") pod "0d2da299-4e48-4135-a749-7a2adf3f7ec9" (UID: "0d2da299-4e48-4135-a749-7a2adf3f7ec9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.916424 4767 scope.go:117] "RemoveContainer" containerID="e2a40d99e8f68af20b6fa59961e265d0c3832a7052b6c49304e606de6ba4d2da" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.940433 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:33 crc kubenswrapper[4767]: I0317 16:04:33.940474 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2da299-4e48-4135-a749-7a2adf3f7ec9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.133674 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-658d967468-8hprg"] Mar 17 16:04:34 crc kubenswrapper[4767]: W0317 16:04:34.147385 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-f646134bf7c3e81d39895e638d392ddf0d379219ae5e27f819f2a8aecd6a8030 WatchSource:0}: Error finding container f646134bf7c3e81d39895e638d392ddf0d379219ae5e27f819f2a8aecd6a8030: Status 404 returned error can't find the container with id f646134bf7c3e81d39895e638d392ddf0d379219ae5e27f819f2a8aecd6a8030 Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.188373 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-66ff489c4c-47nsf"] Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.226550 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-crqkh" event={"ID":"4ad1c7b7-377f-488c-af38-bf30a79c3589","Type":"ContainerDied","Data":"ddc2cee446a82701e0054130b4a12a4a20bbf6e195e503d9fde5759efdb3b2e8"} Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.226607 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddc2cee446a82701e0054130b4a12a4a20bbf6e195e503d9fde5759efdb3b2e8" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.235308 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-658d967468-8hprg" event={"ID":"00a982e7-017e-4f3c-a8aa-21eab359cad4","Type":"ContainerStarted","Data":"f646134bf7c3e81d39895e638d392ddf0d379219ae5e27f819f2a8aecd6a8030"} Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.619239 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.655218 4767 scope.go:117] "RemoveContainer" containerID="dfe0080987b662004c5cc0a001db82ea1328fb2304d1b10898150fe5a5f55213" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.693787 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.714915 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.715505 4767 scope.go:117] "RemoveContainer" containerID="b673dbe3c730b2a2acdf223e799163fe36598b87eb20e0ad865592a2c81c71ac" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.727652 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:04:34 crc kubenswrapper[4767]: E0317 16:04:34.728493 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="proxy-httpd" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.728516 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="proxy-httpd" Mar 17 16:04:34 crc kubenswrapper[4767]: E0317 16:04:34.728536 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="ceilometer-central-agent" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.728551 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="ceilometer-central-agent" Mar 17 16:04:34 crc kubenswrapper[4767]: E0317 16:04:34.728562 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="sg-core" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.728570 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="sg-core" Mar 17 16:04:34 crc kubenswrapper[4767]: E0317 16:04:34.728592 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerName="registry-server" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.728606 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerName="registry-server" Mar 17 16:04:34 crc kubenswrapper[4767]: E0317 16:04:34.728630 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerName="extract-content" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.728640 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerName="extract-content" Mar 17 16:04:34 crc kubenswrapper[4767]: E0317 16:04:34.728688 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerName="extract-utilities" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.728695 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerName="extract-utilities" Mar 17 16:04:34 crc kubenswrapper[4767]: E0317 16:04:34.728719 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="ceilometer-notification-agent" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.728726 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="ceilometer-notification-agent" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.729058 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="ceilometer-notification-agent" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.729076 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="ceilometer-central-agent" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.729093 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="proxy-httpd" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.729110 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="sg-core" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.729135 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ad1c7b7-377f-488c-af38-bf30a79c3589" containerName="registry-server" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.741452 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.751313 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.753465 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckvrd\" (UniqueName: \"kubernetes.io/projected/4ad1c7b7-377f-488c-af38-bf30a79c3589-kube-api-access-ckvrd\") pod \"4ad1c7b7-377f-488c-af38-bf30a79c3589\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.753697 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-catalog-content\") pod \"4ad1c7b7-377f-488c-af38-bf30a79c3589\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.753873 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-utilities\") pod \"4ad1c7b7-377f-488c-af38-bf30a79c3589\" (UID: \"4ad1c7b7-377f-488c-af38-bf30a79c3589\") " Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.754910 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-utilities" (OuterVolumeSpecName: "utilities") pod "4ad1c7b7-377f-488c-af38-bf30a79c3589" (UID: "4ad1c7b7-377f-488c-af38-bf30a79c3589"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.755232 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.755353 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.756566 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.761575 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad1c7b7-377f-488c-af38-bf30a79c3589-kube-api-access-ckvrd" (OuterVolumeSpecName: "kube-api-access-ckvrd") pod "4ad1c7b7-377f-488c-af38-bf30a79c3589" (UID: "4ad1c7b7-377f-488c-af38-bf30a79c3589"). InnerVolumeSpecName "kube-api-access-ckvrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.842987 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ad1c7b7-377f-488c-af38-bf30a79c3589" (UID: "4ad1c7b7-377f-488c-af38-bf30a79c3589"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.861543 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-config-data\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.861749 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-run-httpd\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.861849 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.861929 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvc8d\" (UniqueName: \"kubernetes.io/projected/08838a10-c199-4dc9-bae2-93415ef39ca6-kube-api-access-bvc8d\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.861959 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-scripts\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.862059 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-log-httpd\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.862112 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.862248 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckvrd\" (UniqueName: \"kubernetes.io/projected/4ad1c7b7-377f-488c-af38-bf30a79c3589-kube-api-access-ckvrd\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.862275 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad1c7b7-377f-488c-af38-bf30a79c3589-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.964062 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-log-httpd\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.964133 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.964229 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-config-data\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.964374 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-run-httpd\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.964433 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.964484 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvc8d\" (UniqueName: \"kubernetes.io/projected/08838a10-c199-4dc9-bae2-93415ef39ca6-kube-api-access-bvc8d\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.964509 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-scripts\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.965398 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-run-httpd\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.965814 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-log-httpd\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.976471 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-scripts\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.979906 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.974488 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.992660 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-config-data\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:34 crc kubenswrapper[4767]: I0317 16:04:34.994784 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvc8d\" (UniqueName: \"kubernetes.io/projected/08838a10-c199-4dc9-bae2-93415ef39ca6-kube-api-access-bvc8d\") pod \"ceilometer-0\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " pod="openstack/ceilometer-0" Mar 17 16:04:35 crc kubenswrapper[4767]: I0317 16:04:35.089283 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-q896r"] Mar 17 16:04:35 crc kubenswrapper[4767]: I0317 16:04:35.224348 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:04:35 crc kubenswrapper[4767]: I0317 16:04:35.534293 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" path="/var/lib/kubelet/pods/0d2da299-4e48-4135-a749-7a2adf3f7ec9/volumes" Mar 17 16:04:35 crc kubenswrapper[4767]: I0317 16:04:35.549711 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" event={"ID":"87fb15fd-3cec-4d9a-a316-7f09d9a3062f","Type":"ContainerStarted","Data":"0aa758e93aaf41290d8a31d89a468d95ae7aa16811530b8589d64f9fc539a306"} Mar 17 16:04:35 crc kubenswrapper[4767]: I0317 16:04:35.558486 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-66ff489c4c-47nsf" event={"ID":"2610e145-1410-4330-bcca-808faf035214","Type":"ContainerStarted","Data":"87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7"} Mar 17 16:04:35 crc kubenswrapper[4767]: I0317 16:04:35.558535 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-66ff489c4c-47nsf" event={"ID":"2610e145-1410-4330-bcca-808faf035214","Type":"ContainerStarted","Data":"19efcb8894fdf242e184e7df57c2455f700702871ff1c0f5f0d9be62fa6860f2"} Mar 17 16:04:35 crc kubenswrapper[4767]: I0317 16:04:35.560238 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:35 crc kubenswrapper[4767]: I0317 16:04:35.562249 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-crqkh" Mar 17 16:04:35 crc kubenswrapper[4767]: I0317 16:04:35.599989 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-66ff489c4c-47nsf" podStartSLOduration=16.599966513 podStartE2EDuration="16.599966513s" podCreationTimestamp="2026-03-17 16:04:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:35.59252925 +0000 UTC m=+1667.005845297" watchObservedRunningTime="2026-03-17 16:04:35.599966513 +0000 UTC m=+1667.013282560" Mar 17 16:04:35 crc kubenswrapper[4767]: I0317 16:04:35.661223 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-crqkh"] Mar 17 16:04:35 crc kubenswrapper[4767]: I0317 16:04:35.694711 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-crqkh"] Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.084017 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-9559455fd-vntnw"] Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.110483 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-75777db76d-5z4qz"] Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.156638 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58b6b75dc-pjv6d"] Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.201638 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5bdf466bd4-x9rf5"] Mar 17 16:04:36 crc kubenswrapper[4767]: W0317 16:04:36.342995 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-12e3c38bada1455b32016242209dff4f42ea450915eb2bd79df88618a4bd864a WatchSource:0}: Error finding container 12e3c38bada1455b32016242209dff4f42ea450915eb2bd79df88618a4bd864a: Status 404 returned error can't find the container with id 12e3c38bada1455b32016242209dff4f42ea450915eb2bd79df88618a4bd864a Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.344359 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-56f748dcbf-9x7mb"] Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.386370 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6449cd4787-bslls"] Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.438638 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-56879488c7-5jbhc"] Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.454611 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.579119 4767 generic.go:334] "Generic (PLEG): container finished" podID="87fb15fd-3cec-4d9a-a316-7f09d9a3062f" containerID="b409be9c263934dcad6bec99c14f68c74de2cb7bba7ed96b1cd1dab4ca421d79" exitCode=0 Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.579205 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" event={"ID":"87fb15fd-3cec-4d9a-a316-7f09d9a3062f","Type":"ContainerDied","Data":"b409be9c263934dcad6bec99c14f68c74de2cb7bba7ed96b1cd1dab4ca421d79"} Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.603749 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6449cd4787-bslls" event={"ID":"6fdb5a6c-03a0-4184-8f33-3b7244792cc5","Type":"ContainerStarted","Data":"14fd68462401f4b9e9a32f3322499914628008b9f9298e9a1b777c9d70ad6e84"} Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.607577 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08838a10-c199-4dc9-bae2-93415ef39ca6","Type":"ContainerStarted","Data":"616cc83e4a78acb06cef32fda4bd694ffe10a73cb05346c7f0213a5728220c5c"} Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.611411 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-56879488c7-5jbhc" event={"ID":"65779f05-d74b-49a7-a31a-2de7f9005e59","Type":"ContainerStarted","Data":"11908b2000eead5bd9e1af2f8cd3d7804599246e375750d8483a9f3e90344c1f"} Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.620576 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" event={"ID":"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c","Type":"ContainerStarted","Data":"ee10d09d620425ffc6e31e2cff117403df18c283fb631367bbb5d33b7d5bc463"} Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.623576 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-75777db76d-5z4qz" event={"ID":"3caf16d7-044a-4efc-a3ba-c494255df8ff","Type":"ContainerStarted","Data":"12e3c38bada1455b32016242209dff4f42ea450915eb2bd79df88618a4bd864a"} Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.626034 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9559455fd-vntnw" event={"ID":"33f239ab-85ed-4803-a0db-c31dc5301d28","Type":"ContainerStarted","Data":"10689c0e51f1e6d4f4030b8fd5f677031eaedf4ed42a2d43f5141213dab52fc7"} Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.631367 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5bdf466bd4-x9rf5" event={"ID":"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6","Type":"ContainerStarted","Data":"afba0c2425cec5df41454c84fdd531b5dc53d4cdfd2ba651aa15254f7532244b"} Mar 17 16:04:36 crc kubenswrapper[4767]: I0317 16:04:36.636880 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" event={"ID":"43b3b246-f5d3-4311-ab67-d30e53ac7082","Type":"ContainerStarted","Data":"08e43ff110b79b14aa80a0d973f9c845c33c453a241ce7b8746fd250b01acbd4"} Mar 17 16:04:37 crc kubenswrapper[4767]: I0317 16:04:37.436699 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ad1c7b7-377f-488c-af38-bf30a79c3589" path="/var/lib/kubelet/pods/4ad1c7b7-377f-488c-af38-bf30a79c3589/volumes" Mar 17 16:04:37 crc kubenswrapper[4767]: I0317 16:04:37.664107 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" event={"ID":"87fb15fd-3cec-4d9a-a316-7f09d9a3062f","Type":"ContainerStarted","Data":"87e9dbcc6a20ba9d0d5a91bfc60d63dbb651c0a0512804ea242ef42b5ac9c183"} Mar 17 16:04:37 crc kubenswrapper[4767]: I0317 16:04:37.665767 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:37 crc kubenswrapper[4767]: I0317 16:04:37.669559 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5bdf466bd4-x9rf5" event={"ID":"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6","Type":"ContainerStarted","Data":"c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8"} Mar 17 16:04:37 crc kubenswrapper[4767]: I0317 16:04:37.670210 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:37 crc kubenswrapper[4767]: I0317 16:04:37.678286 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-56879488c7-5jbhc" event={"ID":"65779f05-d74b-49a7-a31a-2de7f9005e59","Type":"ContainerStarted","Data":"60310b999086a1beb533258ea23e383a8571ce64431b108a386658cc08d1b5b9"} Mar 17 16:04:37 crc kubenswrapper[4767]: I0317 16:04:37.714240 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" podStartSLOduration=18.714207653 podStartE2EDuration="18.714207653s" podCreationTimestamp="2026-03-17 16:04:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:37.691428482 +0000 UTC m=+1669.104744539" watchObservedRunningTime="2026-03-17 16:04:37.714207653 +0000 UTC m=+1669.127523700" Mar 17 16:04:37 crc kubenswrapper[4767]: I0317 16:04:37.754182 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5bdf466bd4-x9rf5" podStartSLOduration=8.754143236 podStartE2EDuration="8.754143236s" podCreationTimestamp="2026-03-17 16:04:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:37.726273138 +0000 UTC m=+1669.139589185" watchObservedRunningTime="2026-03-17 16:04:37.754143236 +0000 UTC m=+1669.167459283" Mar 17 16:04:38 crc kubenswrapper[4767]: I0317 16:04:38.705348 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:04:38 crc kubenswrapper[4767]: I0317 16:04:38.707581 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" containerName="glance-log" containerID="cri-o://35c908e5327b918655403169d8b84c8c61ba41ef35f65ece1a57a0e8e323fbab" gracePeriod=30 Mar 17 16:04:38 crc kubenswrapper[4767]: I0317 16:04:38.707782 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" containerName="glance-httpd" containerID="cri-o://a593e4fd0fc22e116160db18ca3009f07986bca775a60abd49ebe1115947d2bf" gracePeriod=30 Mar 17 16:04:39 crc kubenswrapper[4767]: I0317 16:04:39.874037 4767 generic.go:334] "Generic (PLEG): container finished" podID="5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" containerID="35c908e5327b918655403169d8b84c8c61ba41ef35f65ece1a57a0e8e323fbab" exitCode=143 Mar 17 16:04:39 crc kubenswrapper[4767]: I0317 16:04:39.875805 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f","Type":"ContainerDied","Data":"35c908e5327b918655403169d8b84c8c61ba41ef35f65ece1a57a0e8e323fbab"} Mar 17 16:04:40 crc kubenswrapper[4767]: I0317 16:04:40.512832 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.058040 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9559455fd-vntnw" event={"ID":"33f239ab-85ed-4803-a0db-c31dc5301d28","Type":"ContainerStarted","Data":"03ea80cff1aaa98e502d790d0e55b170ea022d03e49e82273cf3a07c30ab1749"} Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.060149 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.069942 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-658d967468-8hprg" event={"ID":"00a982e7-017e-4f3c-a8aa-21eab359cad4","Type":"ContainerStarted","Data":"cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db"} Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.070474 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-658d967468-8hprg" podUID="00a982e7-017e-4f3c-a8aa-21eab359cad4" containerName="heat-cfnapi" containerID="cri-o://cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db" gracePeriod=60 Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.070741 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.091634 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-56879488c7-5jbhc" event={"ID":"65779f05-d74b-49a7-a31a-2de7f9005e59","Type":"ContainerStarted","Data":"e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f"} Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.093494 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.093970 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.106344 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-9559455fd-vntnw" podStartSLOduration=8.961792154 podStartE2EDuration="12.106314813s" podCreationTimestamp="2026-03-17 16:04:29 +0000 UTC" firstStartedPulling="2026-03-17 16:04:36.095510476 +0000 UTC m=+1667.508826523" lastFinishedPulling="2026-03-17 16:04:39.240033135 +0000 UTC m=+1670.653349182" observedRunningTime="2026-03-17 16:04:41.087854965 +0000 UTC m=+1672.501171022" watchObservedRunningTime="2026-03-17 16:04:41.106314813 +0000 UTC m=+1672.519630870" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.122376 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.132209 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.156229 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-658d967468-8hprg" podStartSLOduration=17.062975235 podStartE2EDuration="22.15620224s" podCreationTimestamp="2026-03-17 16:04:19 +0000 UTC" firstStartedPulling="2026-03-17 16:04:34.154044437 +0000 UTC m=+1665.567360484" lastFinishedPulling="2026-03-17 16:04:39.247271442 +0000 UTC m=+1670.660587489" observedRunningTime="2026-03-17 16:04:41.121793256 +0000 UTC m=+1672.535109303" watchObservedRunningTime="2026-03-17 16:04:41.15620224 +0000 UTC m=+1672.569518287" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.161596 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" event={"ID":"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c","Type":"ContainerStarted","Data":"eeee687ced3d03655be88efb84e80b7eb4e725c2cf3e8174bddddb3851ae83e3"} Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.162994 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.200543 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08838a10-c199-4dc9-bae2-93415ef39ca6","Type":"ContainerStarted","Data":"390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678"} Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.258378 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6449cd4787-bslls" podStartSLOduration=6.402004165 podStartE2EDuration="9.258341791s" podCreationTimestamp="2026-03-17 16:04:32 +0000 UTC" firstStartedPulling="2026-03-17 16:04:36.419679998 +0000 UTC m=+1667.832996045" lastFinishedPulling="2026-03-17 16:04:39.276017624 +0000 UTC m=+1670.689333671" observedRunningTime="2026-03-17 16:04:41.227070677 +0000 UTC m=+1672.640386724" watchObservedRunningTime="2026-03-17 16:04:41.258341791 +0000 UTC m=+1672.671657838" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.293936 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-56879488c7-5jbhc" podStartSLOduration=21.293910108 podStartE2EDuration="21.293910108s" podCreationTimestamp="2026-03-17 16:04:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:41.290326816 +0000 UTC m=+1672.703642863" watchObservedRunningTime="2026-03-17 16:04:41.293910108 +0000 UTC m=+1672.707226155" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.365862 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" podStartSLOduration=9.236396608 podStartE2EDuration="12.365823795s" podCreationTimestamp="2026-03-17 16:04:29 +0000 UTC" firstStartedPulling="2026-03-17 16:04:36.149024967 +0000 UTC m=+1667.562341004" lastFinishedPulling="2026-03-17 16:04:39.278452154 +0000 UTC m=+1670.691768191" observedRunningTime="2026-03-17 16:04:41.335788386 +0000 UTC m=+1672.749104453" watchObservedRunningTime="2026-03-17 16:04:41.365823795 +0000 UTC m=+1672.779139852" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.479983 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-2vgbq"] Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.482698 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2vgbq" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.601295 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2vgbq"] Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.603011 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-operator-scripts\") pod \"nova-api-db-create-2vgbq\" (UID: \"24e92570-fb4d-4fa2-ae4f-cc11da8f764a\") " pod="openstack/nova-api-db-create-2vgbq" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.603047 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vd44\" (UniqueName: \"kubernetes.io/projected/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-kube-api-access-4vd44\") pod \"nova-api-db-create-2vgbq\" (UID: \"24e92570-fb4d-4fa2-ae4f-cc11da8f764a\") " pod="openstack/nova-api-db-create-2vgbq" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.880501 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-operator-scripts\") pod \"nova-api-db-create-2vgbq\" (UID: \"24e92570-fb4d-4fa2-ae4f-cc11da8f764a\") " pod="openstack/nova-api-db-create-2vgbq" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.880857 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vd44\" (UniqueName: \"kubernetes.io/projected/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-kube-api-access-4vd44\") pod \"nova-api-db-create-2vgbq\" (UID: \"24e92570-fb4d-4fa2-ae4f-cc11da8f764a\") " pod="openstack/nova-api-db-create-2vgbq" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.882921 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-operator-scripts\") pod \"nova-api-db-create-2vgbq\" (UID: \"24e92570-fb4d-4fa2-ae4f-cc11da8f764a\") " pod="openstack/nova-api-db-create-2vgbq" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.895132 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-gjbh4"] Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.897186 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gjbh4" Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.982877 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-gjbh4"] Mar 17 16:04:41 crc kubenswrapper[4767]: I0317 16:04:41.991927 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vd44\" (UniqueName: \"kubernetes.io/projected/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-kube-api-access-4vd44\") pod \"nova-api-db-create-2vgbq\" (UID: \"24e92570-fb4d-4fa2-ae4f-cc11da8f764a\") " pod="openstack/nova-api-db-create-2vgbq" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.024475 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a6a8fc-9770-4e16-8f14-371b008d3742-operator-scripts\") pod \"nova-cell0-db-create-gjbh4\" (UID: \"d9a6a8fc-9770-4e16-8f14-371b008d3742\") " pod="openstack/nova-cell0-db-create-gjbh4" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.028064 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn5p6\" (UniqueName: \"kubernetes.io/projected/d9a6a8fc-9770-4e16-8f14-371b008d3742-kube-api-access-vn5p6\") pod \"nova-cell0-db-create-gjbh4\" (UID: \"d9a6a8fc-9770-4e16-8f14-371b008d3742\") " pod="openstack/nova-cell0-db-create-gjbh4" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.077132 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.097343 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-f1bf-account-create-update-w5cxr"] Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.099616 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f1bf-account-create-update-w5cxr" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.106564 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.137905 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2vgbq" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.138747 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f1bf-account-create-update-w5cxr"] Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.139281 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efb414e8-1117-4306-85b8-311102e2d8fb-operator-scripts\") pod \"nova-api-f1bf-account-create-update-w5cxr\" (UID: \"efb414e8-1117-4306-85b8-311102e2d8fb\") " pod="openstack/nova-api-f1bf-account-create-update-w5cxr" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.139392 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzk56\" (UniqueName: \"kubernetes.io/projected/efb414e8-1117-4306-85b8-311102e2d8fb-kube-api-access-vzk56\") pod \"nova-api-f1bf-account-create-update-w5cxr\" (UID: \"efb414e8-1117-4306-85b8-311102e2d8fb\") " pod="openstack/nova-api-f1bf-account-create-update-w5cxr" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.139539 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn5p6\" (UniqueName: \"kubernetes.io/projected/d9a6a8fc-9770-4e16-8f14-371b008d3742-kube-api-access-vn5p6\") pod \"nova-cell0-db-create-gjbh4\" (UID: \"d9a6a8fc-9770-4e16-8f14-371b008d3742\") " pod="openstack/nova-cell0-db-create-gjbh4" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.139706 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a6a8fc-9770-4e16-8f14-371b008d3742-operator-scripts\") pod \"nova-cell0-db-create-gjbh4\" (UID: \"d9a6a8fc-9770-4e16-8f14-371b008d3742\") " pod="openstack/nova-cell0-db-create-gjbh4" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.141092 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a6a8fc-9770-4e16-8f14-371b008d3742-operator-scripts\") pod \"nova-cell0-db-create-gjbh4\" (UID: \"d9a6a8fc-9770-4e16-8f14-371b008d3742\") " pod="openstack/nova-cell0-db-create-gjbh4" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.183462 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn5p6\" (UniqueName: \"kubernetes.io/projected/d9a6a8fc-9770-4e16-8f14-371b008d3742-kube-api-access-vn5p6\") pod \"nova-cell0-db-create-gjbh4\" (UID: \"d9a6a8fc-9770-4e16-8f14-371b008d3742\") " pod="openstack/nova-cell0-db-create-gjbh4" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.305000 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efb414e8-1117-4306-85b8-311102e2d8fb-operator-scripts\") pod \"nova-api-f1bf-account-create-update-w5cxr\" (UID: \"efb414e8-1117-4306-85b8-311102e2d8fb\") " pod="openstack/nova-api-f1bf-account-create-update-w5cxr" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.305235 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzk56\" (UniqueName: \"kubernetes.io/projected/efb414e8-1117-4306-85b8-311102e2d8fb-kube-api-access-vzk56\") pod \"nova-api-f1bf-account-create-update-w5cxr\" (UID: \"efb414e8-1117-4306-85b8-311102e2d8fb\") " pod="openstack/nova-api-f1bf-account-create-update-w5cxr" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.306542 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efb414e8-1117-4306-85b8-311102e2d8fb-operator-scripts\") pod \"nova-api-f1bf-account-create-update-w5cxr\" (UID: \"efb414e8-1117-4306-85b8-311102e2d8fb\") " pod="openstack/nova-api-f1bf-account-create-update-w5cxr" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.353274 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gjbh4" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.356884 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:04:42 crc kubenswrapper[4767]: E0317 16:04:42.358431 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.382468 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzk56\" (UniqueName: \"kubernetes.io/projected/efb414e8-1117-4306-85b8-311102e2d8fb-kube-api-access-vzk56\") pod \"nova-api-f1bf-account-create-update-w5cxr\" (UID: \"efb414e8-1117-4306-85b8-311102e2d8fb\") " pod="openstack/nova-api-f1bf-account-create-update-w5cxr" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.440781 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-5zdrh"] Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.452747 4767 generic.go:334] "Generic (PLEG): container finished" podID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" containerID="eeee687ced3d03655be88efb84e80b7eb4e725c2cf3e8174bddddb3851ae83e3" exitCode=1 Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.453677 4767 scope.go:117] "RemoveContainer" containerID="eeee687ced3d03655be88efb84e80b7eb4e725c2cf3e8174bddddb3851ae83e3" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.460324 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-75777db76d-5z4qz" podUID="3caf16d7-044a-4efc-a3ba-c494255df8ff" containerName="heat-api" containerID="cri-o://b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e" gracePeriod=60 Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.476946 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ed4a-account-create-update-49dhk"] Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.477441 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5zdrh" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.477476 4767 generic.go:334] "Generic (PLEG): container finished" podID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerID="03ea80cff1aaa98e502d790d0e55b170ea022d03e49e82273cf3a07c30ab1749" exitCode=1 Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.478607 4767 scope.go:117] "RemoveContainer" containerID="03ea80cff1aaa98e502d790d0e55b170ea022d03e49e82273cf3a07c30ab1749" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.520573 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.520643 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" event={"ID":"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c","Type":"ContainerDied","Data":"eeee687ced3d03655be88efb84e80b7eb4e725c2cf3e8174bddddb3851ae83e3"} Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.520710 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-75777db76d-5z4qz" event={"ID":"3caf16d7-044a-4efc-a3ba-c494255df8ff","Type":"ContainerStarted","Data":"b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e"} Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.520738 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9559455fd-vntnw" event={"ID":"33f239ab-85ed-4803-a0db-c31dc5301d28","Type":"ContainerDied","Data":"03ea80cff1aaa98e502d790d0e55b170ea022d03e49e82273cf3a07c30ab1749"} Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.520876 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.643125 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" event={"ID":"43b3b246-f5d3-4311-ab67-d30e53ac7082","Type":"ContainerStarted","Data":"47cdf9affcddbe0aa7ad601baeab2ee25ee63588a6b2bca9a3c0ee67e20f0c0d"} Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.644464 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.653210 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.653589 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5zdrh"] Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.670791 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6449cd4787-bslls" event={"ID":"6fdb5a6c-03a0-4184-8f33-3b7244792cc5","Type":"ContainerStarted","Data":"e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670"} Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.726975 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ed4a-account-create-update-49dhk"] Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.753096 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwhtv\" (UniqueName: \"kubernetes.io/projected/c1137107-c196-4307-b2c4-90a6e9a0aab4-kube-api-access-wwhtv\") pod \"nova-cell0-ed4a-account-create-update-49dhk\" (UID: \"c1137107-c196-4307-b2c4-90a6e9a0aab4\") " pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.753377 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e8a6ca-f669-4eef-8509-f30f7ee31447-operator-scripts\") pod \"nova-cell1-db-create-5zdrh\" (UID: \"63e8a6ca-f669-4eef-8509-f30f7ee31447\") " pod="openstack/nova-cell1-db-create-5zdrh" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.753577 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1137107-c196-4307-b2c4-90a6e9a0aab4-operator-scripts\") pod \"nova-cell0-ed4a-account-create-update-49dhk\" (UID: \"c1137107-c196-4307-b2c4-90a6e9a0aab4\") " pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.753597 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgvzp\" (UniqueName: \"kubernetes.io/projected/63e8a6ca-f669-4eef-8509-f30f7ee31447-kube-api-access-dgvzp\") pod \"nova-cell1-db-create-5zdrh\" (UID: \"63e8a6ca-f669-4eef-8509-f30f7ee31447\") " pod="openstack/nova-cell1-db-create-5zdrh" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.774935 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-54a5-account-create-update-qzw6x"] Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.779708 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.786065 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.805065 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.858301 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1137107-c196-4307-b2c4-90a6e9a0aab4-operator-scripts\") pod \"nova-cell0-ed4a-account-create-update-49dhk\" (UID: \"c1137107-c196-4307-b2c4-90a6e9a0aab4\") " pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.858365 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgvzp\" (UniqueName: \"kubernetes.io/projected/63e8a6ca-f669-4eef-8509-f30f7ee31447-kube-api-access-dgvzp\") pod \"nova-cell1-db-create-5zdrh\" (UID: \"63e8a6ca-f669-4eef-8509-f30f7ee31447\") " pod="openstack/nova-cell1-db-create-5zdrh" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.858467 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwhtv\" (UniqueName: \"kubernetes.io/projected/c1137107-c196-4307-b2c4-90a6e9a0aab4-kube-api-access-wwhtv\") pod \"nova-cell0-ed4a-account-create-update-49dhk\" (UID: \"c1137107-c196-4307-b2c4-90a6e9a0aab4\") " pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.858608 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e8a6ca-f669-4eef-8509-f30f7ee31447-operator-scripts\") pod \"nova-cell1-db-create-5zdrh\" (UID: \"63e8a6ca-f669-4eef-8509-f30f7ee31447\") " pod="openstack/nova-cell1-db-create-5zdrh" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.875977 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-54a5-account-create-update-qzw6x"] Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.877909 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1137107-c196-4307-b2c4-90a6e9a0aab4-operator-scripts\") pod \"nova-cell0-ed4a-account-create-update-49dhk\" (UID: \"c1137107-c196-4307-b2c4-90a6e9a0aab4\") " pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.903692 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwhtv\" (UniqueName: \"kubernetes.io/projected/c1137107-c196-4307-b2c4-90a6e9a0aab4-kube-api-access-wwhtv\") pod \"nova-cell0-ed4a-account-create-update-49dhk\" (UID: \"c1137107-c196-4307-b2c4-90a6e9a0aab4\") " pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.941153 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgvzp\" (UniqueName: \"kubernetes.io/projected/63e8a6ca-f669-4eef-8509-f30f7ee31447-kube-api-access-dgvzp\") pod \"nova-cell1-db-create-5zdrh\" (UID: \"63e8a6ca-f669-4eef-8509-f30f7ee31447\") " pod="openstack/nova-cell1-db-create-5zdrh" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.944328 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f1bf-account-create-update-w5cxr" Mar 17 16:04:42 crc kubenswrapper[4767]: E0317 16:04:42.947702 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c5ebc5_6ab9_4d98_9a82_0f3c0752906c.slice/crio-conmon-eeee687ced3d03655be88efb84e80b7eb4e725c2cf3e8174bddddb3851ae83e3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c5ebc5_6ab9_4d98_9a82_0f3c0752906c.slice/crio-eeee687ced3d03655be88efb84e80b7eb4e725c2cf3e8174bddddb3851ae83e3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice/crio-03ea80cff1aaa98e502d790d0e55b170ea022d03e49e82273cf3a07c30ab1749.scope\": RecentStats: unable to find data in memory cache]" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.950211 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e8a6ca-f669-4eef-8509-f30f7ee31447-operator-scripts\") pod \"nova-cell1-db-create-5zdrh\" (UID: \"63e8a6ca-f669-4eef-8509-f30f7ee31447\") " pod="openstack/nova-cell1-db-create-5zdrh" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.968055 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/285fe2d6-1df4-4412-a107-a12bcb38e91f-operator-scripts\") pod \"nova-cell1-54a5-account-create-update-qzw6x\" (UID: \"285fe2d6-1df4-4412-a107-a12bcb38e91f\") " pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" Mar 17 16:04:42 crc kubenswrapper[4767]: I0317 16:04:42.968511 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2hjd\" (UniqueName: \"kubernetes.io/projected/285fe2d6-1df4-4412-a107-a12bcb38e91f-kube-api-access-l2hjd\") pod \"nova-cell1-54a5-account-create-update-qzw6x\" (UID: \"285fe2d6-1df4-4412-a107-a12bcb38e91f\") " pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.001453 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5zdrh" Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.053639 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-75777db76d-5z4qz" podStartSLOduration=21.14018685 podStartE2EDuration="24.053608609s" podCreationTimestamp="2026-03-17 16:04:19 +0000 UTC" firstStartedPulling="2026-03-17 16:04:36.363672996 +0000 UTC m=+1667.776989043" lastFinishedPulling="2026-03-17 16:04:39.277094755 +0000 UTC m=+1670.690410802" observedRunningTime="2026-03-17 16:04:42.699499301 +0000 UTC m=+1674.112815358" watchObservedRunningTime="2026-03-17 16:04:43.053608609 +0000 UTC m=+1674.466924676" Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.072663 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/285fe2d6-1df4-4412-a107-a12bcb38e91f-operator-scripts\") pod \"nova-cell1-54a5-account-create-update-qzw6x\" (UID: \"285fe2d6-1df4-4412-a107-a12bcb38e91f\") " pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.072821 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2hjd\" (UniqueName: \"kubernetes.io/projected/285fe2d6-1df4-4412-a107-a12bcb38e91f-kube-api-access-l2hjd\") pod \"nova-cell1-54a5-account-create-update-qzw6x\" (UID: \"285fe2d6-1df4-4412-a107-a12bcb38e91f\") " pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.074156 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/285fe2d6-1df4-4412-a107-a12bcb38e91f-operator-scripts\") pod \"nova-cell1-54a5-account-create-update-qzw6x\" (UID: \"285fe2d6-1df4-4412-a107-a12bcb38e91f\") " pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.076122 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.102505 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" podStartSLOduration=8.239848162 podStartE2EDuration="11.102479317s" podCreationTimestamp="2026-03-17 16:04:32 +0000 UTC" firstStartedPulling="2026-03-17 16:04:36.383440832 +0000 UTC m=+1667.796756879" lastFinishedPulling="2026-03-17 16:04:39.246071987 +0000 UTC m=+1670.659388034" observedRunningTime="2026-03-17 16:04:42.763050209 +0000 UTC m=+1674.176366266" watchObservedRunningTime="2026-03-17 16:04:43.102479317 +0000 UTC m=+1674.515795364" Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.124354 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2hjd\" (UniqueName: \"kubernetes.io/projected/285fe2d6-1df4-4412-a107-a12bcb38e91f-kube-api-access-l2hjd\") pod \"nova-cell1-54a5-account-create-update-qzw6x\" (UID: \"285fe2d6-1df4-4412-a107-a12bcb38e91f\") " pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.133247 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.653946 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-gjbh4"] Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.717892 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2vgbq"] Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.778985 4767 generic.go:334] "Generic (PLEG): container finished" podID="5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" containerID="a593e4fd0fc22e116160db18ca3009f07986bca775a60abd49ebe1115947d2bf" exitCode=0 Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.779330 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f","Type":"ContainerDied","Data":"a593e4fd0fc22e116160db18ca3009f07986bca775a60abd49ebe1115947d2bf"} Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.794704 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" event={"ID":"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c","Type":"ContainerStarted","Data":"c7d4dd32069ee1072975d0601b1ffa8e62d738ddbb5023ad4b3f2b088b3efcdf"} Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.795554 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:43 crc kubenswrapper[4767]: I0317 16:04:43.826383 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08838a10-c199-4dc9-bae2-93415ef39ca6","Type":"ContainerStarted","Data":"0bd1b41c99c5fee3341231a56e77f3f105618713dca3db21ef95535978c355eb"} Mar 17 16:04:43 crc kubenswrapper[4767]: E0317 16:04:43.947892 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.016585 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.536255 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.745373 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-internal-tls-certs\") pod \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.745516 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-logs\") pod \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.745707 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-combined-ca-bundle\") pod \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.747280 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-logs" (OuterVolumeSpecName: "logs") pod "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" (UID: "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.747665 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.747712 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-config-data\") pod \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.747764 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-httpd-run\") pod \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.747911 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhlhg\" (UniqueName: \"kubernetes.io/projected/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-kube-api-access-xhlhg\") pod \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.747970 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-scripts\") pod \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.749109 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.750148 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" (UID: "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.778962 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-scripts" (OuterVolumeSpecName: "scripts") pod "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" (UID: "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.808660 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-kube-api-access-xhlhg" (OuterVolumeSpecName: "kube-api-access-xhlhg") pod "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" (UID: "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f"). InnerVolumeSpecName "kube-api-access-xhlhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.847626 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2vgbq" event={"ID":"24e92570-fb4d-4fa2-ae4f-cc11da8f764a","Type":"ContainerStarted","Data":"9c389228536150f157fedb29d96e58bb0fcdd2211a266a61a40f255ba8e1aa01"} Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.858636 4767 generic.go:334] "Generic (PLEG): container finished" podID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" containerID="c7d4dd32069ee1072975d0601b1ffa8e62d738ddbb5023ad4b3f2b088b3efcdf" exitCode=1 Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.858747 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" event={"ID":"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c","Type":"ContainerDied","Data":"c7d4dd32069ee1072975d0601b1ffa8e62d738ddbb5023ad4b3f2b088b3efcdf"} Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.858789 4767 scope.go:117] "RemoveContainer" containerID="eeee687ced3d03655be88efb84e80b7eb4e725c2cf3e8174bddddb3851ae83e3" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.864018 4767 scope.go:117] "RemoveContainer" containerID="c7d4dd32069ee1072975d0601b1ffa8e62d738ddbb5023ad4b3f2b088b3efcdf" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.864045 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gjbh4" event={"ID":"d9a6a8fc-9770-4e16-8f14-371b008d3742","Type":"ContainerStarted","Data":"898ed8832f6b941b5bb20f9c1244dc76126c914f33a93cb74a5a34c4407a8018"} Mar 17 16:04:44 crc kubenswrapper[4767]: E0317 16:04:44.867273 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58b6b75dc-pjv6d_openstack(c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c)\"" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" podUID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.885125 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9559455fd-vntnw" event={"ID":"33f239ab-85ed-4803-a0db-c31dc5301d28","Type":"ContainerStarted","Data":"6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066"} Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.887712 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.907141 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhlhg\" (UniqueName: \"kubernetes.io/projected/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-kube-api-access-xhlhg\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.907221 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:44 crc kubenswrapper[4767]: I0317 16:04:44.907238 4767 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.045679 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" (UID: "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.046566 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f","Type":"ContainerDied","Data":"24af3b22057bd463148d003486e027125a2c4b49bf6932f4acd657274a8cdbdf"} Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.047053 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.048532 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.123871 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.137435 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.147509 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" (UID: "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.190466 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.227670 4767 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:45 crc kubenswrapper[4767]: E0317 16:04:45.246875 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581 podName:5a02e83b-1e68-446f-9fd7-9cb6fb125a6f nodeName:}" failed. No retries permitted until 2026-03-17 16:04:45.746825598 +0000 UTC m=+1677.160141645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "glance" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581") pod "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" (UID: "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f") : kubernetes.io/csi: Unmounter.TearDownAt failed: rpc error: code = Unknown desc = check target path: could not get consistent content of /proc/mounts after 3 attempts Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.265866 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-h4pcz"] Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.266141 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" podUID="2f4160a8-203a-41e6-a90f-f0800ae776b1" containerName="dnsmasq-dns" containerID="cri-o://7b426e927de42d1e528bab380f9497633286b54dc26035261cb9ceeff790bb07" gracePeriod=10 Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.310598 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-config-data" (OuterVolumeSpecName: "config-data") pod "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" (UID: "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.343001 4767 scope.go:117] "RemoveContainer" containerID="a593e4fd0fc22e116160db18ca3009f07986bca775a60abd49ebe1115947d2bf" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.345398 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.441136 4767 scope.go:117] "RemoveContainer" containerID="35c908e5327b918655403169d8b84c8c61ba41ef35f65ece1a57a0e8e323fbab" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.809885 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\" (UID: \"5a02e83b-1e68-446f-9fd7-9cb6fb125a6f\") " Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.833899 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-54a5-account-create-update-qzw6x"] Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.877020 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f1bf-account-create-update-w5cxr"] Mar 17 16:04:45 crc kubenswrapper[4767]: W0317 16:04:45.908134 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-a9f3df5fc86345edfe138aefdfbaeeb838ca491f039c7d110fb7e6384f38eb9e WatchSource:0}: Error finding container a9f3df5fc86345edfe138aefdfbaeeb838ca491f039c7d110fb7e6384f38eb9e: Status 404 returned error can't find the container with id a9f3df5fc86345edfe138aefdfbaeeb838ca491f039c7d110fb7e6384f38eb9e Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.926332 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ed4a-account-create-update-49dhk"] Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.948832 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5zdrh"] Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.965460 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.965512 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:04:45 crc kubenswrapper[4767]: W0317 16:04:45.990474 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1137107_c196_4307_b2c4_90a6e9a0aab4.slice/crio-59308e52d31fae1391cba5905c2021853fa4ca3b518419461424e331d001360a WatchSource:0}: Error finding container 59308e52d31fae1391cba5905c2021853fa4ca3b518419461424e331d001360a: Status 404 returned error can't find the container with id 59308e52d31fae1391cba5905c2021853fa4ca3b518419461424e331d001360a Mar 17 16:04:45 crc kubenswrapper[4767]: I0317 16:04:45.991316 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581" (OuterVolumeSpecName: "glance") pod "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" (UID: "5a02e83b-1e68-446f-9fd7-9cb6fb125a6f"). InnerVolumeSpecName "pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.025292 4767 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") on node \"crc\" " Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.143153 4767 generic.go:334] "Generic (PLEG): container finished" podID="2f4160a8-203a-41e6-a90f-f0800ae776b1" containerID="7b426e927de42d1e528bab380f9497633286b54dc26035261cb9ceeff790bb07" exitCode=0 Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.143243 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" event={"ID":"2f4160a8-203a-41e6-a90f-f0800ae776b1","Type":"ContainerDied","Data":"7b426e927de42d1e528bab380f9497633286b54dc26035261cb9ceeff790bb07"} Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.145145 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gjbh4" event={"ID":"d9a6a8fc-9770-4e16-8f14-371b008d3742","Type":"ContainerStarted","Data":"ece2197718375ad2e2b7f83d4da1945e4eacd599465e12964e926042e8ab2db8"} Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.153759 4767 generic.go:334] "Generic (PLEG): container finished" podID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerID="6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066" exitCode=1 Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.153896 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9559455fd-vntnw" event={"ID":"33f239ab-85ed-4803-a0db-c31dc5301d28","Type":"ContainerDied","Data":"6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066"} Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.155102 4767 scope.go:117] "RemoveContainer" containerID="6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066" Mar 17 16:04:46 crc kubenswrapper[4767]: E0317 16:04:46.155451 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-9559455fd-vntnw_openstack(33f239ab-85ed-4803-a0db-c31dc5301d28)\"" pod="openstack/heat-api-9559455fd-vntnw" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.168080 4767 scope.go:117] "RemoveContainer" containerID="03ea80cff1aaa98e502d790d0e55b170ea022d03e49e82273cf3a07c30ab1749" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.169872 4767 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.170087 4767 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581") on node "crc" Mar 17 16:04:46 crc kubenswrapper[4767]: E0317 16:04:46.184436 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35c908e5327b918655403169d8b84c8c61ba41ef35f65ece1a57a0e8e323fbab\": container with ID starting with 35c908e5327b918655403169d8b84c8c61ba41ef35f65ece1a57a0e8e323fbab not found: ID does not exist" containerID="35c908e5327b918655403169d8b84c8c61ba41ef35f65ece1a57a0e8e323fbab" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.209574 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-gjbh4" podStartSLOduration=5.209541263 podStartE2EDuration="5.209541263s" podCreationTimestamp="2026-03-17 16:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:46.182784948 +0000 UTC m=+1677.596100995" watchObservedRunningTime="2026-03-17 16:04:46.209541263 +0000 UTC m=+1677.622857310" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.215682 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" event={"ID":"c1137107-c196-4307-b2c4-90a6e9a0aab4","Type":"ContainerStarted","Data":"59308e52d31fae1391cba5905c2021853fa4ca3b518419461424e331d001360a"} Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.220401 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.233066 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.248022 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" event={"ID":"285fe2d6-1df4-4412-a107-a12bcb38e91f","Type":"ContainerStarted","Data":"9047fd2850cafa2664014a183b86d91d0b386c26afaebcf2407a7e02440b02bb"} Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.257482 4767 reconciler_common.go:293] "Volume detached for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.285616 4767 scope.go:117] "RemoveContainer" containerID="c7d4dd32069ee1072975d0601b1ffa8e62d738ddbb5023ad4b3f2b088b3efcdf" Mar 17 16:04:46 crc kubenswrapper[4767]: E0317 16:04:46.286312 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58b6b75dc-pjv6d_openstack(c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c)\"" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" podUID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.304276 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:04:46 crc kubenswrapper[4767]: E0317 16:04:46.305421 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" containerName="glance-httpd" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.305446 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" containerName="glance-httpd" Mar 17 16:04:46 crc kubenswrapper[4767]: E0317 16:04:46.305518 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" containerName="glance-log" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.305529 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" containerName="glance-log" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.305875 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" containerName="glance-log" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.306407 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" containerName="glance-httpd" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.308664 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.314271 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.314644 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.371994 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08838a10-c199-4dc9-bae2-93415ef39ca6","Type":"ContainerStarted","Data":"19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c"} Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.442533 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5zdrh" event={"ID":"63e8a6ca-f669-4eef-8509-f30f7ee31447","Type":"ContainerStarted","Data":"a9f3df5fc86345edfe138aefdfbaeeb838ca491f039c7d110fb7e6384f38eb9e"} Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.446488 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f1bf-account-create-update-w5cxr" event={"ID":"efb414e8-1117-4306-85b8-311102e2d8fb","Type":"ContainerStarted","Data":"e4f1fc5d421dc2bd9c32944f148850e3c7906c603d6614c995b39964a2b95272"} Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.640672 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2vgbq" event={"ID":"24e92570-fb4d-4fa2-ae4f-cc11da8f764a","Type":"ContainerStarted","Data":"0729e2cd2713c04092f2dd54f9c0acd5be84ae27edf51f2f8ef07b34198bf8bd"} Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.717429 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.718277 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5951f0d9-dd9d-4b63-9708-370be10fc5dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.718355 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.718498 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4467\" (UniqueName: \"kubernetes.io/projected/5951f0d9-dd9d-4b63-9708-370be10fc5dc-kube-api-access-n4467\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.718566 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5951f0d9-dd9d-4b63-9708-370be10fc5dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.718602 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.718738 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.718875 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.775315 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.825816 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.826003 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5951f0d9-dd9d-4b63-9708-370be10fc5dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.826057 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.826138 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4467\" (UniqueName: \"kubernetes.io/projected/5951f0d9-dd9d-4b63-9708-370be10fc5dc-kube-api-access-n4467\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.827354 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5951f0d9-dd9d-4b63-9708-370be10fc5dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.827469 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.827523 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.827563 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.827723 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5951f0d9-dd9d-4b63-9708-370be10fc5dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.828900 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5951f0d9-dd9d-4b63-9708-370be10fc5dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.850255 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.850307 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5f7c8528f2df23f78ebd936b9a4e66a32c7cd4122d28a5f082f69cd7529229c4/globalmount\"" pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.863787 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.873678 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.892492 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.893154 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5951f0d9-dd9d-4b63-9708-370be10fc5dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:46 crc kubenswrapper[4767]: I0317 16:04:46.921992 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4467\" (UniqueName: \"kubernetes.io/projected/5951f0d9-dd9d-4b63-9708-370be10fc5dc-kube-api-access-n4467\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.074816 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-2vgbq" podStartSLOduration=6.074791141 podStartE2EDuration="6.074791141s" podCreationTimestamp="2026-03-17 16:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:46.740654324 +0000 UTC m=+1678.153970401" watchObservedRunningTime="2026-03-17 16:04:47.074791141 +0000 UTC m=+1678.488107188" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.137965 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2b78ee-ed38-46a3-9d30-2f8f67911581\") pod \"glance-default-internal-api-0\" (UID: \"5951f0d9-dd9d-4b63-9708-370be10fc5dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.212431 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.275091 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.361401 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-nb\") pod \"2f4160a8-203a-41e6-a90f-f0800ae776b1\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.361540 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-svc\") pod \"2f4160a8-203a-41e6-a90f-f0800ae776b1\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.361664 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-sb\") pod \"2f4160a8-203a-41e6-a90f-f0800ae776b1\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.361704 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np485\" (UniqueName: \"kubernetes.io/projected/2f4160a8-203a-41e6-a90f-f0800ae776b1-kube-api-access-np485\") pod \"2f4160a8-203a-41e6-a90f-f0800ae776b1\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.361755 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-swift-storage-0\") pod \"2f4160a8-203a-41e6-a90f-f0800ae776b1\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.361900 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-config\") pod \"2f4160a8-203a-41e6-a90f-f0800ae776b1\" (UID: \"2f4160a8-203a-41e6-a90f-f0800ae776b1\") " Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.406039 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f4160a8-203a-41e6-a90f-f0800ae776b1-kube-api-access-np485" (OuterVolumeSpecName: "kube-api-access-np485") pod "2f4160a8-203a-41e6-a90f-f0800ae776b1" (UID: "2f4160a8-203a-41e6-a90f-f0800ae776b1"). InnerVolumeSpecName "kube-api-access-np485". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.428972 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a02e83b-1e68-446f-9fd7-9cb6fb125a6f" path="/var/lib/kubelet/pods/5a02e83b-1e68-446f-9fd7-9cb6fb125a6f/volumes" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.472267 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np485\" (UniqueName: \"kubernetes.io/projected/2f4160a8-203a-41e6-a90f-f0800ae776b1-kube-api-access-np485\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.544452 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2f4160a8-203a-41e6-a90f-f0800ae776b1" (UID: "2f4160a8-203a-41e6-a90f-f0800ae776b1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.579745 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.584231 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2f4160a8-203a-41e6-a90f-f0800ae776b1" (UID: "2f4160a8-203a-41e6-a90f-f0800ae776b1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.585647 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2f4160a8-203a-41e6-a90f-f0800ae776b1" (UID: "2f4160a8-203a-41e6-a90f-f0800ae776b1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.780146 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.780201 4767 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.844525 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2f4160a8-203a-41e6-a90f-f0800ae776b1" (UID: "2f4160a8-203a-41e6-a90f-f0800ae776b1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.896140 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.903370 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9kh2v"] Mar 17 16:04:47 crc kubenswrapper[4767]: E0317 16:04:47.904144 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f4160a8-203a-41e6-a90f-f0800ae776b1" containerName="init" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.904161 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f4160a8-203a-41e6-a90f-f0800ae776b1" containerName="init" Mar 17 16:04:47 crc kubenswrapper[4767]: E0317 16:04:47.904210 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f4160a8-203a-41e6-a90f-f0800ae776b1" containerName="dnsmasq-dns" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.904218 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f4160a8-203a-41e6-a90f-f0800ae776b1" containerName="dnsmasq-dns" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.904489 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f4160a8-203a-41e6-a90f-f0800ae776b1" containerName="dnsmasq-dns" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.906477 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.908785 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" event={"ID":"2f4160a8-203a-41e6-a90f-f0800ae776b1","Type":"ContainerDied","Data":"89469eed3c5666995f50f5309cf2b253b49b11c04e7890fe945935f52393344d"} Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.908842 4767 scope.go:117] "RemoveContainer" containerID="7b426e927de42d1e528bab380f9497633286b54dc26035261cb9ceeff790bb07" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.908956 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.919234 4767 generic.go:334] "Generic (PLEG): container finished" podID="d9a6a8fc-9770-4e16-8f14-371b008d3742" containerID="ece2197718375ad2e2b7f83d4da1945e4eacd599465e12964e926042e8ab2db8" exitCode=0 Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.919311 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gjbh4" event={"ID":"d9a6a8fc-9770-4e16-8f14-371b008d3742","Type":"ContainerDied","Data":"ece2197718375ad2e2b7f83d4da1945e4eacd599465e12964e926042e8ab2db8"} Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.927449 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-config" (OuterVolumeSpecName: "config") pod "2f4160a8-203a-41e6-a90f-f0800ae776b1" (UID: "2f4160a8-203a-41e6-a90f-f0800ae776b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.938770 4767 scope.go:117] "RemoveContainer" containerID="6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066" Mar 17 16:04:47 crc kubenswrapper[4767]: E0317 16:04:47.939280 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-9559455fd-vntnw_openstack(33f239ab-85ed-4803-a0db-c31dc5301d28)\"" pod="openstack/heat-api-9559455fd-vntnw" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.940431 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9kh2v"] Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.959291 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f1bf-account-create-update-w5cxr" event={"ID":"efb414e8-1117-4306-85b8-311102e2d8fb","Type":"ContainerStarted","Data":"234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c"} Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.964567 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.964665 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.966116 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="proxy-httpd" containerStatusID={"Type":"cri-o","ID":"60310b999086a1beb533258ea23e383a8571ce64431b108a386658cc08d1b5b9"} pod="openstack/swift-proxy-56879488c7-5jbhc" containerMessage="Container proxy-httpd failed liveness probe, will be restarted" Mar 17 16:04:47 crc kubenswrapper[4767]: I0317 16:04:47.966211 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" containerID="cri-o://60310b999086a1beb533258ea23e383a8571ce64431b108a386658cc08d1b5b9" gracePeriod=30 Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:47.998046 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxjdt\" (UniqueName: \"kubernetes.io/projected/65a045e9-219c-4432-9d80-09a944c85013-kube-api-access-nxjdt\") pod \"certified-operators-9kh2v\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:47.998394 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-catalog-content\") pod \"certified-operators-9kh2v\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:47.998473 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-utilities\") pod \"certified-operators-9kh2v\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:47.998589 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f4160a8-203a-41e6-a90f-f0800ae776b1-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.029846 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.067856 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.068040 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.079609 4767 generic.go:334] "Generic (PLEG): container finished" podID="24e92570-fb4d-4fa2-ae4f-cc11da8f764a" containerID="0729e2cd2713c04092f2dd54f9c0acd5be84ae27edf51f2f8ef07b34198bf8bd" exitCode=0 Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.079762 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2vgbq" event={"ID":"24e92570-fb4d-4fa2-ae4f-cc11da8f764a","Type":"ContainerDied","Data":"0729e2cd2713c04092f2dd54f9c0acd5be84ae27edf51f2f8ef07b34198bf8bd"} Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.096725 4767 scope.go:117] "RemoveContainer" containerID="c7d4dd32069ee1072975d0601b1ffa8e62d738ddbb5023ad4b3f2b088b3efcdf" Mar 17 16:04:48 crc kubenswrapper[4767]: E0317 16:04:48.097018 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58b6b75dc-pjv6d_openstack(c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c)\"" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" podUID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.097675 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" event={"ID":"285fe2d6-1df4-4412-a107-a12bcb38e91f","Type":"ContainerStarted","Data":"72df4826ccc3d3f9a8847830042fbbc37df929e953b73a207af394e476ca5a43"} Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.102398 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-catalog-content\") pod \"certified-operators-9kh2v\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.102523 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-utilities\") pod \"certified-operators-9kh2v\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.102630 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxjdt\" (UniqueName: \"kubernetes.io/projected/65a045e9-219c-4432-9d80-09a944c85013-kube-api-access-nxjdt\") pod \"certified-operators-9kh2v\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.103115 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-catalog-content\") pod \"certified-operators-9kh2v\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.103404 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-utilities\") pod \"certified-operators-9kh2v\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.164773 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-f1bf-account-create-update-w5cxr" podStartSLOduration=7.164744485 podStartE2EDuration="7.164744485s" podCreationTimestamp="2026-03-17 16:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:48.101666961 +0000 UTC m=+1679.514983008" watchObservedRunningTime="2026-03-17 16:04:48.164744485 +0000 UTC m=+1679.578060532" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.177483 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxjdt\" (UniqueName: \"kubernetes.io/projected/65a045e9-219c-4432-9d80-09a944c85013-kube-api-access-nxjdt\") pod \"certified-operators-9kh2v\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:04:48 crc kubenswrapper[4767]: I0317 16:04:48.260070 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" podStartSLOduration=6.2600439 podStartE2EDuration="6.2600439s" podCreationTimestamp="2026-03-17 16:04:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:48.134264493 +0000 UTC m=+1679.547580530" watchObservedRunningTime="2026-03-17 16:04:48.2600439 +0000 UTC m=+1679.673359947" Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:49.159821 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:49.179132 4767 generic.go:334] "Generic (PLEG): container finished" podID="285fe2d6-1df4-4412-a107-a12bcb38e91f" containerID="72df4826ccc3d3f9a8847830042fbbc37df929e953b73a207af394e476ca5a43" exitCode=0 Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:49.179278 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" event={"ID":"285fe2d6-1df4-4412-a107-a12bcb38e91f","Type":"ContainerDied","Data":"72df4826ccc3d3f9a8847830042fbbc37df929e953b73a207af394e476ca5a43"} Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:49.215532 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" event={"ID":"c1137107-c196-4307-b2c4-90a6e9a0aab4","Type":"ContainerStarted","Data":"30521a8308aa007e0c56096559f18a5842284b1cb62c937699d69295f3e22e08"} Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:49.233132 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5zdrh" event={"ID":"63e8a6ca-f669-4eef-8509-f30f7ee31447","Type":"ContainerStarted","Data":"71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94"} Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:49.252883 4767 generic.go:334] "Generic (PLEG): container finished" podID="efb414e8-1117-4306-85b8-311102e2d8fb" containerID="234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c" exitCode=0 Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:49.252943 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f1bf-account-create-update-w5cxr" event={"ID":"efb414e8-1117-4306-85b8-311102e2d8fb","Type":"ContainerDied","Data":"234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c"} Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:49.268219 4767 generic.go:334] "Generic (PLEG): container finished" podID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerID="60310b999086a1beb533258ea23e383a8571ce64431b108a386658cc08d1b5b9" exitCode=0 Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:49.268462 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-56879488c7-5jbhc" event={"ID":"65779f05-d74b-49a7-a31a-2de7f9005e59","Type":"ContainerDied","Data":"60310b999086a1beb533258ea23e383a8571ce64431b108a386658cc08d1b5b9"} Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:49.359482 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" podStartSLOduration=7.359457676 podStartE2EDuration="7.359457676s" podCreationTimestamp="2026-03-17 16:04:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:49.279844369 +0000 UTC m=+1680.693160416" watchObservedRunningTime="2026-03-17 16:04:49.359457676 +0000 UTC m=+1680.772773713" Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:49.859552 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-5zdrh" podStartSLOduration=7.859519108 podStartE2EDuration="7.859519108s" podCreationTimestamp="2026-03-17 16:04:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:49.33898372 +0000 UTC m=+1680.752299777" watchObservedRunningTime="2026-03-17 16:04:49.859519108 +0000 UTC m=+1681.272835155" Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:50.319771 4767 generic.go:334] "Generic (PLEG): container finished" podID="c1137107-c196-4307-b2c4-90a6e9a0aab4" containerID="30521a8308aa007e0c56096559f18a5842284b1cb62c937699d69295f3e22e08" exitCode=0 Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:50.328698 4767 generic.go:334] "Generic (PLEG): container finished" podID="63e8a6ca-f669-4eef-8509-f30f7ee31447" containerID="71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94" exitCode=0 Mar 17 16:04:51 crc kubenswrapper[4767]: E0317 16:04:50.806272 4767 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.453s" Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:50.806628 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gjbh4" event={"ID":"d9a6a8fc-9770-4e16-8f14-371b008d3742","Type":"ContainerDied","Data":"898ed8832f6b941b5bb20f9c1244dc76126c914f33a93cb74a5a34c4407a8018"} Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:50.806660 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="898ed8832f6b941b5bb20f9c1244dc76126c914f33a93cb74a5a34c4407a8018" Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:50.806679 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:50.806760 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" event={"ID":"c1137107-c196-4307-b2c4-90a6e9a0aab4","Type":"ContainerDied","Data":"30521a8308aa007e0c56096559f18a5842284b1cb62c937699d69295f3e22e08"} Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:50.806780 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5zdrh" event={"ID":"63e8a6ca-f669-4eef-8509-f30f7ee31447","Type":"ContainerDied","Data":"71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94"} Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:50.956050 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.226:8080/healthcheck\": dial tcp 10.217.0.226:8080: connect: connection refused" Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:50.956387 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.226:8080/healthcheck\": dial tcp 10.217.0.226:8080: connect: connection refused" Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:50.956434 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.226:8080/healthcheck\": dial tcp 10.217.0.226:8080: connect: connection refused" Mar 17 16:04:51 crc kubenswrapper[4767]: I0317 16:04:51.151232 4767 scope.go:117] "RemoveContainer" containerID="6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066" Mar 17 16:04:51 crc kubenswrapper[4767]: E0317 16:04:51.151783 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-9559455fd-vntnw_openstack(33f239ab-85ed-4803-a0db-c31dc5301d28)\"" pod="openstack/heat-api-9559455fd-vntnw" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" Mar 17 16:04:53 crc kubenswrapper[4767]: E0317 16:04:52.819912 4767 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.465s" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:52.830876 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:52.831025 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5951f0d9-dd9d-4b63-9708-370be10fc5dc","Type":"ContainerStarted","Data":"2b4bf9a7944e9bfec96c97a025f678fa17ac2f81733a0cbd795b33148dcff5ad"} Mar 17 16:04:53 crc kubenswrapper[4767]: E0317 16:04:52.831022 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:52.831080 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:04:53 crc kubenswrapper[4767]: E0317 16:04:52.883405 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-conmon-60310b999086a1beb533258ea23e383a8571ce64431b108a386658cc08d1b5b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1137107_c196_4307_b2c4_90a6e9a0aab4.slice/crio-30521a8308aa007e0c56096559f18a5842284b1cb62c937699d69295f3e22e08.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fe2d6_1df4_4412_a107_a12bcb38e91f.slice/crio-72df4826ccc3d3f9a8847830042fbbc37df929e953b73a207af394e476ca5a43.scope\": RecentStats: unable to find data in memory cache]" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.099087 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-66ff489c4c-47nsf"] Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.438510 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" event={"ID":"c1137107-c196-4307-b2c4-90a6e9a0aab4","Type":"ContainerDied","Data":"59308e52d31fae1391cba5905c2021853fa4ca3b518419461424e331d001360a"} Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.438564 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59308e52d31fae1391cba5905c2021853fa4ca3b518419461424e331d001360a" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.458783 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5zdrh" event={"ID":"63e8a6ca-f669-4eef-8509-f30f7ee31447","Type":"ContainerDied","Data":"a9f3df5fc86345edfe138aefdfbaeeb838ca491f039c7d110fb7e6384f38eb9e"} Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.458863 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9f3df5fc86345edfe138aefdfbaeeb838ca491f039c7d110fb7e6384f38eb9e" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.466135 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f1bf-account-create-update-w5cxr" event={"ID":"efb414e8-1117-4306-85b8-311102e2d8fb","Type":"ContainerDied","Data":"e4f1fc5d421dc2bd9c32944f148850e3c7906c603d6614c995b39964a2b95272"} Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.466236 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4f1fc5d421dc2bd9c32944f148850e3c7906c603d6614c995b39964a2b95272" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.477907 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2vgbq" event={"ID":"24e92570-fb4d-4fa2-ae4f-cc11da8f764a","Type":"ContainerDied","Data":"9c389228536150f157fedb29d96e58bb0fcdd2211a266a61a40f255ba8e1aa01"} Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.477956 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c389228536150f157fedb29d96e58bb0fcdd2211a266a61a40f255ba8e1aa01" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.484006 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-66ff489c4c-47nsf" podUID="2610e145-1410-4330-bcca-808faf035214" containerName="heat-engine" containerID="cri-o://87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7" gracePeriod=60 Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.484388 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" event={"ID":"285fe2d6-1df4-4412-a107-a12bcb38e91f","Type":"ContainerDied","Data":"9047fd2850cafa2664014a183b86d91d0b386c26afaebcf2407a7e02440b02bb"} Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.484417 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9047fd2850cafa2664014a183b86d91d0b386c26afaebcf2407a7e02440b02bb" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.611773 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.622276 4767 scope.go:117] "RemoveContainer" containerID="41fd51b01636d01492ed5e2e39508da2604df8f281a4571a0ad83f67237944a3" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.810900 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gjbh4" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.818670 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.887551 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.893460 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a6a8fc-9770-4e16-8f14-371b008d3742-operator-scripts\") pod \"d9a6a8fc-9770-4e16-8f14-371b008d3742\" (UID: \"d9a6a8fc-9770-4e16-8f14-371b008d3742\") " Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.893941 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn5p6\" (UniqueName: \"kubernetes.io/projected/d9a6a8fc-9770-4e16-8f14-371b008d3742-kube-api-access-vn5p6\") pod \"d9a6a8fc-9770-4e16-8f14-371b008d3742\" (UID: \"d9a6a8fc-9770-4e16-8f14-371b008d3742\") " Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.894736 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9a6a8fc-9770-4e16-8f14-371b008d3742-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d9a6a8fc-9770-4e16-8f14-371b008d3742" (UID: "d9a6a8fc-9770-4e16-8f14-371b008d3742"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.924417 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2vgbq" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.944474 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.226:8080/healthcheck\": dial tcp 10.217.0.226:8080: connect: connection refused" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.944564 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.944590 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9a6a8fc-9770-4e16-8f14-371b008d3742-kube-api-access-vn5p6" (OuterVolumeSpecName: "kube-api-access-vn5p6") pod "d9a6a8fc-9770-4e16-8f14-371b008d3742" (UID: "d9a6a8fc-9770-4e16-8f14-371b008d3742"). InnerVolumeSpecName "kube-api-access-vn5p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.997304 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2hjd\" (UniqueName: \"kubernetes.io/projected/285fe2d6-1df4-4412-a107-a12bcb38e91f-kube-api-access-l2hjd\") pod \"285fe2d6-1df4-4412-a107-a12bcb38e91f\" (UID: \"285fe2d6-1df4-4412-a107-a12bcb38e91f\") " Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.997654 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/285fe2d6-1df4-4412-a107-a12bcb38e91f-operator-scripts\") pod \"285fe2d6-1df4-4412-a107-a12bcb38e91f\" (UID: \"285fe2d6-1df4-4412-a107-a12bcb38e91f\") " Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.998664 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn5p6\" (UniqueName: \"kubernetes.io/projected/d9a6a8fc-9770-4e16-8f14-371b008d3742-kube-api-access-vn5p6\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:53 crc kubenswrapper[4767]: I0317 16:04:53.998697 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a6a8fc-9770-4e16-8f14-371b008d3742-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.001300 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/285fe2d6-1df4-4412-a107-a12bcb38e91f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "285fe2d6-1df4-4412-a107-a12bcb38e91f" (UID: "285fe2d6-1df4-4412-a107-a12bcb38e91f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.123453 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vd44\" (UniqueName: \"kubernetes.io/projected/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-kube-api-access-4vd44\") pod \"24e92570-fb4d-4fa2-ae4f-cc11da8f764a\" (UID: \"24e92570-fb4d-4fa2-ae4f-cc11da8f764a\") " Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.265206 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-operator-scripts\") pod \"24e92570-fb4d-4fa2-ae4f-cc11da8f764a\" (UID: \"24e92570-fb4d-4fa2-ae4f-cc11da8f764a\") " Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.267129 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/285fe2d6-1df4-4412-a107-a12bcb38e91f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.268846 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "24e92570-fb4d-4fa2-ae4f-cc11da8f764a" (UID: "24e92570-fb4d-4fa2-ae4f-cc11da8f764a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.291760 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-kube-api-access-4vd44" (OuterVolumeSpecName: "kube-api-access-4vd44") pod "24e92570-fb4d-4fa2-ae4f-cc11da8f764a" (UID: "24e92570-fb4d-4fa2-ae4f-cc11da8f764a"). InnerVolumeSpecName "kube-api-access-4vd44". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.292657 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/285fe2d6-1df4-4412-a107-a12bcb38e91f-kube-api-access-l2hjd" (OuterVolumeSpecName: "kube-api-access-l2hjd") pod "285fe2d6-1df4-4412-a107-a12bcb38e91f" (UID: "285fe2d6-1df4-4412-a107-a12bcb38e91f"). InnerVolumeSpecName "kube-api-access-l2hjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.358257 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:04:54 crc kubenswrapper[4767]: E0317 16:04:54.358977 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.370649 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2hjd\" (UniqueName: \"kubernetes.io/projected/285fe2d6-1df4-4412-a107-a12bcb38e91f-kube-api-access-l2hjd\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.370684 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vd44\" (UniqueName: \"kubernetes.io/projected/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-kube-api-access-4vd44\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.370695 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24e92570-fb4d-4fa2-ae4f-cc11da8f764a-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.541968 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2vgbq" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.542884 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gjbh4" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.542855 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54a5-account-create-update-qzw6x" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.721482 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f1bf-account-create-update-w5cxr" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.747475 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.782063 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5zdrh" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.890444 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efb414e8-1117-4306-85b8-311102e2d8fb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "efb414e8-1117-4306-85b8-311102e2d8fb" (UID: "efb414e8-1117-4306-85b8-311102e2d8fb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.883901 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efb414e8-1117-4306-85b8-311102e2d8fb-operator-scripts\") pod \"efb414e8-1117-4306-85b8-311102e2d8fb\" (UID: \"efb414e8-1117-4306-85b8-311102e2d8fb\") " Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.890885 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgvzp\" (UniqueName: \"kubernetes.io/projected/63e8a6ca-f669-4eef-8509-f30f7ee31447-kube-api-access-dgvzp\") pod \"63e8a6ca-f669-4eef-8509-f30f7ee31447\" (UID: \"63e8a6ca-f669-4eef-8509-f30f7ee31447\") " Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.891038 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzk56\" (UniqueName: \"kubernetes.io/projected/efb414e8-1117-4306-85b8-311102e2d8fb-kube-api-access-vzk56\") pod \"efb414e8-1117-4306-85b8-311102e2d8fb\" (UID: \"efb414e8-1117-4306-85b8-311102e2d8fb\") " Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.891084 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwhtv\" (UniqueName: \"kubernetes.io/projected/c1137107-c196-4307-b2c4-90a6e9a0aab4-kube-api-access-wwhtv\") pod \"c1137107-c196-4307-b2c4-90a6e9a0aab4\" (UID: \"c1137107-c196-4307-b2c4-90a6e9a0aab4\") " Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.891148 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1137107-c196-4307-b2c4-90a6e9a0aab4-operator-scripts\") pod \"c1137107-c196-4307-b2c4-90a6e9a0aab4\" (UID: \"c1137107-c196-4307-b2c4-90a6e9a0aab4\") " Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.891329 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e8a6ca-f669-4eef-8509-f30f7ee31447-operator-scripts\") pod \"63e8a6ca-f669-4eef-8509-f30f7ee31447\" (UID: \"63e8a6ca-f669-4eef-8509-f30f7ee31447\") " Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.893094 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efb414e8-1117-4306-85b8-311102e2d8fb-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.893625 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e8a6ca-f669-4eef-8509-f30f7ee31447-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "63e8a6ca-f669-4eef-8509-f30f7ee31447" (UID: "63e8a6ca-f669-4eef-8509-f30f7ee31447"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.894018 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1137107-c196-4307-b2c4-90a6e9a0aab4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c1137107-c196-4307-b2c4-90a6e9a0aab4" (UID: "c1137107-c196-4307-b2c4-90a6e9a0aab4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.904637 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1137107-c196-4307-b2c4-90a6e9a0aab4-kube-api-access-wwhtv" (OuterVolumeSpecName: "kube-api-access-wwhtv") pod "c1137107-c196-4307-b2c4-90a6e9a0aab4" (UID: "c1137107-c196-4307-b2c4-90a6e9a0aab4"). InnerVolumeSpecName "kube-api-access-wwhtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.906600 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efb414e8-1117-4306-85b8-311102e2d8fb-kube-api-access-vzk56" (OuterVolumeSpecName: "kube-api-access-vzk56") pod "efb414e8-1117-4306-85b8-311102e2d8fb" (UID: "efb414e8-1117-4306-85b8-311102e2d8fb"). InnerVolumeSpecName "kube-api-access-vzk56". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.922945 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63e8a6ca-f669-4eef-8509-f30f7ee31447-kube-api-access-dgvzp" (OuterVolumeSpecName: "kube-api-access-dgvzp") pod "63e8a6ca-f669-4eef-8509-f30f7ee31447" (UID: "63e8a6ca-f669-4eef-8509-f30f7ee31447"). InnerVolumeSpecName "kube-api-access-dgvzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.983986 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9kh2v"] Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.998468 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzk56\" (UniqueName: \"kubernetes.io/projected/efb414e8-1117-4306-85b8-311102e2d8fb-kube-api-access-vzk56\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.998543 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwhtv\" (UniqueName: \"kubernetes.io/projected/c1137107-c196-4307-b2c4-90a6e9a0aab4-kube-api-access-wwhtv\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.998557 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1137107-c196-4307-b2c4-90a6e9a0aab4-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.998569 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e8a6ca-f669-4eef-8509-f30f7ee31447-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:54 crc kubenswrapper[4767]: I0317 16:04:54.998580 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgvzp\" (UniqueName: \"kubernetes.io/projected/63e8a6ca-f669-4eef-8509-f30f7ee31447-kube-api-access-dgvzp\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.041648 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.602120 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5951f0d9-dd9d-4b63-9708-370be10fc5dc","Type":"ContainerStarted","Data":"4914fe569dbacced3984e264da1bf5ba9478b9afe752c7a27313da381d1d78aa"} Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.609147 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"049b9f3b-c6db-4274-9f45-c5cac0c0a17b","Type":"ContainerStarted","Data":"62a0bee25c03eb1f4c3402d1f585c9595c55e18e70e38b688e27a7ea34e9fca5"} Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.631668 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-56879488c7-5jbhc" event={"ID":"65779f05-d74b-49a7-a31a-2de7f9005e59","Type":"ContainerStarted","Data":"261b0abb962deabe8287bcb3503cf1a3449b444c9e13a72eaa319e2295b6b18b"} Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.638191 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.655999 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="proxy-server" containerStatusID={"Type":"cri-o","ID":"e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f"} pod="openstack/swift-proxy-56879488c7-5jbhc" containerMessage="Container proxy-server failed liveness probe, will be restarted" Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.656084 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-server" containerID="cri-o://e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f" gracePeriod=30 Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.688986 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=11.753117956 podStartE2EDuration="54.688962359s" podCreationTimestamp="2026-03-17 16:04:01 +0000 UTC" firstStartedPulling="2026-03-17 16:04:04.479758266 +0000 UTC m=+1635.893074313" lastFinishedPulling="2026-03-17 16:04:47.415602669 +0000 UTC m=+1678.828918716" observedRunningTime="2026-03-17 16:04:55.633090801 +0000 UTC m=+1687.046406868" watchObservedRunningTime="2026-03-17 16:04:55.688962359 +0000 UTC m=+1687.102278406" Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.733143 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08838a10-c199-4dc9-bae2-93415ef39ca6","Type":"ContainerStarted","Data":"786f37d75313e4409a3b52ab9c2aef5ca0cb3e9ad79f97230e225bdf1b660666"} Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.733574 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="ceilometer-central-agent" containerID="cri-o://390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678" gracePeriod=30 Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.734147 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.734215 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="sg-core" containerID="cri-o://19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c" gracePeriod=30 Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.734364 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="proxy-httpd" containerID="cri-o://786f37d75313e4409a3b52ab9c2aef5ca0cb3e9ad79f97230e225bdf1b660666" gracePeriod=30 Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.737554 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="ceilometer-notification-agent" containerID="cri-o://0bd1b41c99c5fee3341231a56e77f3f105618713dca3db21ef95535978c355eb" gracePeriod=30 Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.771841 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5zdrh" Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.774350 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kh2v" event={"ID":"65a045e9-219c-4432-9d80-09a944c85013","Type":"ContainerStarted","Data":"1d5e03d5f46a6b457764cbf327336b3881a4b232d151d5ab3221aaaacf8fdfec"} Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.774473 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.774568 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f1bf-account-create-update-w5cxr" Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.779798 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ed4a-account-create-update-49dhk" Mar 17 16:04:55 crc kubenswrapper[4767]: I0317 16:04:55.862219 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=10.303775606 podStartE2EDuration="21.862194384s" podCreationTimestamp="2026-03-17 16:04:34 +0000 UTC" firstStartedPulling="2026-03-17 16:04:36.510397883 +0000 UTC m=+1667.923713930" lastFinishedPulling="2026-03-17 16:04:48.068816671 +0000 UTC m=+1679.482132708" observedRunningTime="2026-03-17 16:04:55.855034299 +0000 UTC m=+1687.268350366" watchObservedRunningTime="2026-03-17 16:04:55.862194384 +0000 UTC m=+1687.275510431" Mar 17 16:04:56 crc kubenswrapper[4767]: I0317 16:04:56.134317 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:04:56 crc kubenswrapper[4767]: I0317 16:04:56.745476 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-58b6b75dc-pjv6d"] Mar 17 16:04:56 crc kubenswrapper[4767]: I0317 16:04:56.860038 4767 generic.go:334] "Generic (PLEG): container finished" podID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerID="19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c" exitCode=2 Mar 17 16:04:56 crc kubenswrapper[4767]: I0317 16:04:56.860342 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08838a10-c199-4dc9-bae2-93415ef39ca6","Type":"ContainerDied","Data":"19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c"} Mar 17 16:04:56 crc kubenswrapper[4767]: I0317 16:04:56.863419 4767 generic.go:334] "Generic (PLEG): container finished" podID="65a045e9-219c-4432-9d80-09a944c85013" containerID="3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191" exitCode=0 Mar 17 16:04:56 crc kubenswrapper[4767]: I0317 16:04:56.863449 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kh2v" event={"ID":"65a045e9-219c-4432-9d80-09a944c85013","Type":"ContainerDied","Data":"3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191"} Mar 17 16:04:56 crc kubenswrapper[4767]: I0317 16:04:56.909093 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:57.977748 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" event={"ID":"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c","Type":"ContainerDied","Data":"ee10d09d620425ffc6e31e2cff117403df18c283fb631367bbb5d33b7d5bc463"} Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:57.978219 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee10d09d620425ffc6e31e2cff117403df18c283fb631367bbb5d33b7d5bc463" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.023377 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dsn5f"] Mar 17 16:04:58 crc kubenswrapper[4767]: E0317 16:04:58.023991 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1137107-c196-4307-b2c4-90a6e9a0aab4" containerName="mariadb-account-create-update" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024005 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1137107-c196-4307-b2c4-90a6e9a0aab4" containerName="mariadb-account-create-update" Mar 17 16:04:58 crc kubenswrapper[4767]: E0317 16:04:58.024024 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9a6a8fc-9770-4e16-8f14-371b008d3742" containerName="mariadb-database-create" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024030 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9a6a8fc-9770-4e16-8f14-371b008d3742" containerName="mariadb-database-create" Mar 17 16:04:58 crc kubenswrapper[4767]: E0317 16:04:58.024048 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24e92570-fb4d-4fa2-ae4f-cc11da8f764a" containerName="mariadb-database-create" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024054 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="24e92570-fb4d-4fa2-ae4f-cc11da8f764a" containerName="mariadb-database-create" Mar 17 16:04:58 crc kubenswrapper[4767]: E0317 16:04:58.024068 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="285fe2d6-1df4-4412-a107-a12bcb38e91f" containerName="mariadb-account-create-update" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024074 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="285fe2d6-1df4-4412-a107-a12bcb38e91f" containerName="mariadb-account-create-update" Mar 17 16:04:58 crc kubenswrapper[4767]: E0317 16:04:58.024084 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e8a6ca-f669-4eef-8509-f30f7ee31447" containerName="mariadb-database-create" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024091 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e8a6ca-f669-4eef-8509-f30f7ee31447" containerName="mariadb-database-create" Mar 17 16:04:58 crc kubenswrapper[4767]: E0317 16:04:58.024153 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efb414e8-1117-4306-85b8-311102e2d8fb" containerName="mariadb-account-create-update" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024162 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="efb414e8-1117-4306-85b8-311102e2d8fb" containerName="mariadb-account-create-update" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024440 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="24e92570-fb4d-4fa2-ae4f-cc11da8f764a" containerName="mariadb-database-create" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024463 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="efb414e8-1117-4306-85b8-311102e2d8fb" containerName="mariadb-account-create-update" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024476 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9a6a8fc-9770-4e16-8f14-371b008d3742" containerName="mariadb-database-create" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024485 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1137107-c196-4307-b2c4-90a6e9a0aab4" containerName="mariadb-account-create-update" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024496 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="285fe2d6-1df4-4412-a107-a12bcb38e91f" containerName="mariadb-account-create-update" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.024507 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="63e8a6ca-f669-4eef-8509-f30f7ee31447" containerName="mariadb-database-create" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.025495 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.058824 4767 generic.go:334] "Generic (PLEG): container finished" podID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerID="0bd1b41c99c5fee3341231a56e77f3f105618713dca3db21ef95535978c355eb" exitCode=0 Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.058917 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08838a10-c199-4dc9-bae2-93415ef39ca6","Type":"ContainerDied","Data":"0bd1b41c99c5fee3341231a56e77f3f105618713dca3db21ef95535978c355eb"} Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.060553 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-kp9tf" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.060845 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.061236 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.314655 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-config-data\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.314786 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggltx\" (UniqueName: \"kubernetes.io/projected/dba94162-c0cc-46e7-a4ca-45836a003a08-kube-api-access-ggltx\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.314813 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.314922 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-scripts\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.417767 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-config-data\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.423455 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggltx\" (UniqueName: \"kubernetes.io/projected/dba94162-c0cc-46e7-a4ca-45836a003a08-kube-api-access-ggltx\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.423511 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.423762 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-scripts\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.457591 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-config-data\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.458082 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-scripts\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.476010 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggltx\" (UniqueName: \"kubernetes.io/projected/dba94162-c0cc-46e7-a4ca-45836a003a08-kube-api-access-ggltx\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.479823 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dsn5f\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.499835 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.557893 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dsn5f"] Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.565730 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-6449cd4787-bslls" podUID="6fdb5a6c-03a0-4184-8f33-3b7244792cc5" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.230:8004/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.674245 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95z8k\" (UniqueName: \"kubernetes.io/projected/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-kube-api-access-95z8k\") pod \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.674378 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data-custom\") pod \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.674532 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-combined-ca-bundle\") pod \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.674598 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data\") pod \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\" (UID: \"c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c\") " Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.683123 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" (UID: "c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.685651 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.693577 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.731465 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-kube-api-access-95z8k" (OuterVolumeSpecName: "kube-api-access-95z8k") pod "c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" (UID: "c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c"). InnerVolumeSpecName "kube-api-access-95z8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.791423 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95z8k\" (UniqueName: \"kubernetes.io/projected/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-kube-api-access-95z8k\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.845369 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" (UID: "c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.894800 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:58 crc kubenswrapper[4767]: I0317 16:04:58.928473 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data" (OuterVolumeSpecName: "config-data") pod "c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" (UID: "c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:04:59 crc kubenswrapper[4767]: I0317 16:04:59.282115 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:04:59 crc kubenswrapper[4767]: I0317 16:04:59.424476 4767 generic.go:334] "Generic (PLEG): container finished" podID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerID="e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f" exitCode=0 Mar 17 16:04:59 crc kubenswrapper[4767]: I0317 16:04:59.550284 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58b6b75dc-pjv6d" Mar 17 16:04:59 crc kubenswrapper[4767]: I0317 16:04:59.561436 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-56879488c7-5jbhc" event={"ID":"65779f05-d74b-49a7-a31a-2de7f9005e59","Type":"ContainerDied","Data":"e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f"} Mar 17 16:04:59 crc kubenswrapper[4767]: I0317 16:04:59.561495 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kh2v" event={"ID":"65a045e9-219c-4432-9d80-09a944c85013","Type":"ContainerStarted","Data":"f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004"} Mar 17 16:04:59 crc kubenswrapper[4767]: I0317 16:04:59.561511 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5951f0d9-dd9d-4b63-9708-370be10fc5dc","Type":"ContainerStarted","Data":"1a2c0b82a177ec7cc38c5a9368fc962155cb6d9ef21c76da0a57d08da429fb39"} Mar 17 16:04:59 crc kubenswrapper[4767]: I0317 16:04:59.655751 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=13.655730165 podStartE2EDuration="13.655730165s" podCreationTimestamp="2026-03-17 16:04:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:04:59.632377577 +0000 UTC m=+1691.045693634" watchObservedRunningTime="2026-03-17 16:04:59.655730165 +0000 UTC m=+1691.069046212" Mar 17 16:04:59 crc kubenswrapper[4767]: I0317 16:04:59.688601 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0d2da299-4e48-4135-a749-7a2adf3f7ec9" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.217:3000/\": dial tcp 10.217.0.217:3000: i/o timeout (Client.Timeout exceeded while awaiting headers)" Mar 17 16:05:00 crc kubenswrapper[4767]: E0317 16:04:59.866373 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 17 16:05:00 crc kubenswrapper[4767]: E0317 16:04:59.874625 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 17 16:05:00 crc kubenswrapper[4767]: E0317 16:04:59.881402 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 17 16:05:00 crc kubenswrapper[4767]: E0317 16:04:59.881545 4767 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-66ff489c4c-47nsf" podUID="2610e145-1410-4330-bcca-808faf035214" containerName="heat-engine" Mar 17 16:05:00 crc kubenswrapper[4767]: I0317 16:04:59.964514 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:05:00 crc kubenswrapper[4767]: I0317 16:04:59.980692 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dsn5f"] Mar 17 16:05:00 crc kubenswrapper[4767]: I0317 16:05:00.591763 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-56879488c7-5jbhc" event={"ID":"65779f05-d74b-49a7-a31a-2de7f9005e59","Type":"ContainerStarted","Data":"8f33a43e4c8b7395faa1d69634ff639b3f18d35ee887c6396aab3353f4876683"} Mar 17 16:05:00 crc kubenswrapper[4767]: I0317 16:05:00.594673 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:05:00 crc kubenswrapper[4767]: I0317 16:05:00.622039 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:05:00 crc kubenswrapper[4767]: I0317 16:05:00.815605 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-58b6b75dc-pjv6d"] Mar 17 16:05:00 crc kubenswrapper[4767]: I0317 16:05:00.815682 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-58b6b75dc-pjv6d"] Mar 17 16:05:00 crc kubenswrapper[4767]: I0317 16:05:00.945617 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:05:01 crc kubenswrapper[4767]: I0317 16:05:01.429624 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" path="/var/lib/kubelet/pods/c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c/volumes" Mar 17 16:05:01 crc kubenswrapper[4767]: I0317 16:05:01.642278 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dsn5f" event={"ID":"dba94162-c0cc-46e7-a4ca-45836a003a08","Type":"ContainerStarted","Data":"0cfe300fc7f0e3a19680132d190529f6b8a6c4da47e533665e133efe3d1cd005"} Mar 17 16:05:01 crc kubenswrapper[4767]: I0317 16:05:01.663076 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:05:02 crc kubenswrapper[4767]: I0317 16:05:02.909316 4767 generic.go:334] "Generic (PLEG): container finished" podID="65a045e9-219c-4432-9d80-09a944c85013" containerID="f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004" exitCode=0 Mar 17 16:05:02 crc kubenswrapper[4767]: I0317 16:05:02.911656 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kh2v" event={"ID":"65a045e9-219c-4432-9d80-09a944c85013","Type":"ContainerDied","Data":"f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004"} Mar 17 16:05:03 crc kubenswrapper[4767]: I0317 16:05:03.356672 4767 scope.go:117] "RemoveContainer" containerID="6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066" Mar 17 16:05:03 crc kubenswrapper[4767]: I0317 16:05:03.786732 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:05:04 crc kubenswrapper[4767]: I0317 16:05:04.195306 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:05:04 crc kubenswrapper[4767]: I0317 16:05:04.332040 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-9559455fd-vntnw"] Mar 17 16:05:04 crc kubenswrapper[4767]: I0317 16:05:04.394979 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-56879488c7-5jbhc" Mar 17 16:05:05 crc kubenswrapper[4767]: I0317 16:05:05.346308 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:05:05 crc kubenswrapper[4767]: I0317 16:05:05.434698 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9559455fd-vntnw" event={"ID":"33f239ab-85ed-4803-a0db-c31dc5301d28","Type":"ContainerStarted","Data":"6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d"} Mar 17 16:05:05 crc kubenswrapper[4767]: I0317 16:05:05.435235 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-9559455fd-vntnw" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerName="heat-api" containerID="cri-o://6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d" gracePeriod=60 Mar 17 16:05:05 crc kubenswrapper[4767]: I0317 16:05:05.435495 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:05:05 crc kubenswrapper[4767]: I0317 16:05:05.495420 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kh2v" event={"ID":"65a045e9-219c-4432-9d80-09a944c85013","Type":"ContainerStarted","Data":"70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41"} Mar 17 16:05:05 crc kubenswrapper[4767]: I0317 16:05:05.559690 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9kh2v" podStartSLOduration=11.814722342 podStartE2EDuration="18.559646607s" podCreationTimestamp="2026-03-17 16:04:47 +0000 UTC" firstStartedPulling="2026-03-17 16:04:56.867465497 +0000 UTC m=+1688.280781534" lastFinishedPulling="2026-03-17 16:05:03.612389752 +0000 UTC m=+1695.025705799" observedRunningTime="2026-03-17 16:05:05.517419339 +0000 UTC m=+1696.930735386" watchObservedRunningTime="2026-03-17 16:05:05.559646607 +0000 UTC m=+1696.972962654" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.352720 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.355727 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:05:06 crc kubenswrapper[4767]: E0317 16:05:06.356311 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.460587 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-combined-ca-bundle\") pod \"33f239ab-85ed-4803-a0db-c31dc5301d28\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.460801 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9rfw\" (UniqueName: \"kubernetes.io/projected/33f239ab-85ed-4803-a0db-c31dc5301d28-kube-api-access-t9rfw\") pod \"33f239ab-85ed-4803-a0db-c31dc5301d28\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.461126 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data\") pod \"33f239ab-85ed-4803-a0db-c31dc5301d28\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.461218 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data-custom\") pod \"33f239ab-85ed-4803-a0db-c31dc5301d28\" (UID: \"33f239ab-85ed-4803-a0db-c31dc5301d28\") " Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.493403 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "33f239ab-85ed-4803-a0db-c31dc5301d28" (UID: "33f239ab-85ed-4803-a0db-c31dc5301d28"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.508570 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33f239ab-85ed-4803-a0db-c31dc5301d28-kube-api-access-t9rfw" (OuterVolumeSpecName: "kube-api-access-t9rfw") pod "33f239ab-85ed-4803-a0db-c31dc5301d28" (UID: "33f239ab-85ed-4803-a0db-c31dc5301d28"). InnerVolumeSpecName "kube-api-access-t9rfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.541121 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33f239ab-85ed-4803-a0db-c31dc5301d28" (UID: "33f239ab-85ed-4803-a0db-c31dc5301d28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.571134 4767 generic.go:334] "Generic (PLEG): container finished" podID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerID="6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d" exitCode=1 Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.571216 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9559455fd-vntnw" event={"ID":"33f239ab-85ed-4803-a0db-c31dc5301d28","Type":"ContainerDied","Data":"6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d"} Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.571255 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9559455fd-vntnw" event={"ID":"33f239ab-85ed-4803-a0db-c31dc5301d28","Type":"ContainerDied","Data":"10689c0e51f1e6d4f4030b8fd5f677031eaedf4ed42a2d43f5141213dab52fc7"} Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.571279 4767 scope.go:117] "RemoveContainer" containerID="6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.571499 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-9559455fd-vntnw" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.587065 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.587104 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9rfw\" (UniqueName: \"kubernetes.io/projected/33f239ab-85ed-4803-a0db-c31dc5301d28-kube-api-access-t9rfw\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.587119 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.590368 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data" (OuterVolumeSpecName: "config-data") pod "33f239ab-85ed-4803-a0db-c31dc5301d28" (UID: "33f239ab-85ed-4803-a0db-c31dc5301d28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:06 crc kubenswrapper[4767]: I0317 16:05:06.948657 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33f239ab-85ed-4803-a0db-c31dc5301d28-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.034834 4767 scope.go:117] "RemoveContainer" containerID="6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.062533 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-9559455fd-vntnw"] Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.092662 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-9559455fd-vntnw"] Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.169620 4767 scope.go:117] "RemoveContainer" containerID="6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d" Mar 17 16:05:07 crc kubenswrapper[4767]: E0317 16:05:07.177531 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d\": container with ID starting with 6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d not found: ID does not exist" containerID="6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.177606 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d"} err="failed to get container status \"6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d\": rpc error: code = NotFound desc = could not find container \"6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d\": container with ID starting with 6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d not found: ID does not exist" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.177647 4767 scope.go:117] "RemoveContainer" containerID="6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066" Mar 17 16:05:07 crc kubenswrapper[4767]: E0317 16:05:07.179442 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066\": container with ID starting with 6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066 not found: ID does not exist" containerID="6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.179471 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066"} err="failed to get container status \"6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066\": rpc error: code = NotFound desc = could not find container \"6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066\": container with ID starting with 6db07a5a78e872b5e4309cca1c26b7ea4a32bde9eae57a1c8b21039ccad0a066 not found: ID does not exist" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.213315 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.214947 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.388629 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" path="/var/lib/kubelet/pods/33f239ab-85ed-4803-a0db-c31dc5301d28/volumes" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.396233 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.404963 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.591975 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 17 16:05:07 crc kubenswrapper[4767]: I0317 16:05:07.592022 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.327308 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.413296 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data-custom\") pod \"2610e145-1410-4330-bcca-808faf035214\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.413451 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gshrj\" (UniqueName: \"kubernetes.io/projected/2610e145-1410-4330-bcca-808faf035214-kube-api-access-gshrj\") pod \"2610e145-1410-4330-bcca-808faf035214\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.413491 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-combined-ca-bundle\") pod \"2610e145-1410-4330-bcca-808faf035214\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.413657 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data\") pod \"2610e145-1410-4330-bcca-808faf035214\" (UID: \"2610e145-1410-4330-bcca-808faf035214\") " Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.420921 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2610e145-1410-4330-bcca-808faf035214-kube-api-access-gshrj" (OuterVolumeSpecName: "kube-api-access-gshrj") pod "2610e145-1410-4330-bcca-808faf035214" (UID: "2610e145-1410-4330-bcca-808faf035214"). InnerVolumeSpecName "kube-api-access-gshrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.448467 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2610e145-1410-4330-bcca-808faf035214" (UID: "2610e145-1410-4330-bcca-808faf035214"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.510948 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2610e145-1410-4330-bcca-808faf035214" (UID: "2610e145-1410-4330-bcca-808faf035214"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.518373 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.518425 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gshrj\" (UniqueName: \"kubernetes.io/projected/2610e145-1410-4330-bcca-808faf035214-kube-api-access-gshrj\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.518442 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.577069 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data" (OuterVolumeSpecName: "config-data") pod "2610e145-1410-4330-bcca-808faf035214" (UID: "2610e145-1410-4330-bcca-808faf035214"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.622184 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2610e145-1410-4330-bcca-808faf035214-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.623603 4767 generic.go:334] "Generic (PLEG): container finished" podID="2610e145-1410-4330-bcca-808faf035214" containerID="87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7" exitCode=0 Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.623796 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-66ff489c4c-47nsf" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.623801 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-66ff489c4c-47nsf" event={"ID":"2610e145-1410-4330-bcca-808faf035214","Type":"ContainerDied","Data":"87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7"} Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.623864 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-66ff489c4c-47nsf" event={"ID":"2610e145-1410-4330-bcca-808faf035214","Type":"ContainerDied","Data":"19efcb8894fdf242e184e7df57c2455f700702871ff1c0f5f0d9be62fa6860f2"} Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.623884 4767 scope.go:117] "RemoveContainer" containerID="87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7" Mar 17 16:05:08 crc kubenswrapper[4767]: I0317 16:05:08.702352 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-66ff489c4c-47nsf"] Mar 17 16:05:09 crc kubenswrapper[4767]: I0317 16:05:09.005253 4767 scope.go:117] "RemoveContainer" containerID="87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7" Mar 17 16:05:09 crc kubenswrapper[4767]: I0317 16:05:09.026480 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-66ff489c4c-47nsf"] Mar 17 16:05:09 crc kubenswrapper[4767]: E0317 16:05:09.054973 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7\": container with ID starting with 87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7 not found: ID does not exist" containerID="87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7" Mar 17 16:05:09 crc kubenswrapper[4767]: I0317 16:05:09.055612 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7"} err="failed to get container status \"87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7\": rpc error: code = NotFound desc = could not find container \"87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7\": container with ID starting with 87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7 not found: ID does not exist" Mar 17 16:05:09 crc kubenswrapper[4767]: I0317 16:05:09.419434 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2610e145-1410-4330-bcca-808faf035214" path="/var/lib/kubelet/pods/2610e145-1410-4330-bcca-808faf035214/volumes" Mar 17 16:05:09 crc kubenswrapper[4767]: I0317 16:05:09.420741 4767 scope.go:117] "RemoveContainer" containerID="20330680ac4fb421b4d9350049f827cc802f164445428f995497f7a928a3b552" Mar 17 16:05:10 crc kubenswrapper[4767]: I0317 16:05:10.008782 4767 generic.go:334] "Generic (PLEG): container finished" podID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerID="390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678" exitCode=0 Mar 17 16:05:10 crc kubenswrapper[4767]: I0317 16:05:10.008850 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08838a10-c199-4dc9-bae2-93415ef39ca6","Type":"ContainerDied","Data":"390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678"} Mar 17 16:05:11 crc kubenswrapper[4767]: I0317 16:05:11.463953 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:05:11 crc kubenswrapper[4767]: I0317 16:05:11.469645 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="302b036a-e22c-4c7d-a924-aa2aa150f81a" containerName="glance-log" containerID="cri-o://f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156" gracePeriod=30 Mar 17 16:05:11 crc kubenswrapper[4767]: I0317 16:05:11.470091 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="302b036a-e22c-4c7d-a924-aa2aa150f81a" containerName="glance-httpd" containerID="cri-o://02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e" gracePeriod=30 Mar 17 16:05:13 crc kubenswrapper[4767]: I0317 16:05:13.153659 4767 generic.go:334] "Generic (PLEG): container finished" podID="302b036a-e22c-4c7d-a924-aa2aa150f81a" containerID="f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156" exitCode=143 Mar 17 16:05:13 crc kubenswrapper[4767]: I0317 16:05:13.154345 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"302b036a-e22c-4c7d-a924-aa2aa150f81a","Type":"ContainerDied","Data":"f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156"} Mar 17 16:05:13 crc kubenswrapper[4767]: I0317 16:05:13.819731 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:05:13 crc kubenswrapper[4767]: I0317 16:05:13.820155 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:05:14 crc kubenswrapper[4767]: I0317 16:05:14.893303 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9kh2v" podUID="65a045e9-219c-4432-9d80-09a944c85013" containerName="registry-server" probeResult="failure" output=< Mar 17 16:05:14 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:05:14 crc kubenswrapper[4767]: > Mar 17 16:05:15 crc kubenswrapper[4767]: I0317 16:05:15.397939 4767 patch_prober.go:28] interesting pod/logging-loki-query-frontend-ff66c4dc9-4jtkt container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 16:05:15 crc kubenswrapper[4767]: I0317 16:05:15.398394 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" podUID="ce1ebb9b-1558-4190-9e47-4ef6ba3160da" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 16:05:15 crc kubenswrapper[4767]: I0317 16:05:15.398127 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" podUID="8bfc70f3-ead2-43cc-9e90-7df32804d9ac" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 16:05:15 crc kubenswrapper[4767]: I0317 16:05:15.397939 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" podUID="8bfc70f3-ead2-43cc-9e90-7df32804d9ac" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 16:05:16 crc kubenswrapper[4767]: I0317 16:05:16.785304 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 17 16:05:16 crc kubenswrapper[4767]: I0317 16:05:16.785502 4767 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 16:05:16 crc kubenswrapper[4767]: I0317 16:05:16.879159 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 17 16:05:17 crc kubenswrapper[4767]: I0317 16:05:17.361981 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:05:17 crc kubenswrapper[4767]: E0317 16:05:17.364101 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:05:17 crc kubenswrapper[4767]: I0317 16:05:17.594784 4767 generic.go:334] "Generic (PLEG): container finished" podID="302b036a-e22c-4c7d-a924-aa2aa150f81a" containerID="02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e" exitCode=0 Mar 17 16:05:17 crc kubenswrapper[4767]: I0317 16:05:17.595906 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"302b036a-e22c-4c7d-a924-aa2aa150f81a","Type":"ContainerDied","Data":"02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e"} Mar 17 16:05:23 crc kubenswrapper[4767]: I0317 16:05:23.671366 4767 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod2f4160a8-203a-41e6-a90f-f0800ae776b1"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod2f4160a8-203a-41e6-a90f-f0800ae776b1] : Timed out while waiting for systemd to remove kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice" Mar 17 16:05:23 crc kubenswrapper[4767]: E0317 16:05:23.671947 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod2f4160a8-203a-41e6-a90f-f0800ae776b1] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod2f4160a8-203a-41e6-a90f-f0800ae776b1] : Timed out while waiting for systemd to remove kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" podUID="2f4160a8-203a-41e6-a90f-f0800ae776b1" Mar 17 16:05:23 crc kubenswrapper[4767]: I0317 16:05:23.672029 4767 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod2f4160a8-203a-41e6-a90f-f0800ae776b1"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod2f4160a8-203a-41e6-a90f-f0800ae776b1] : Timed out while waiting for systemd to remove kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice" Mar 17 16:05:23 crc kubenswrapper[4767]: I0317 16:05:23.681604 4767 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod2f4160a8-203a-41e6-a90f-f0800ae776b1"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod2f4160a8-203a-41e6-a90f-f0800ae776b1] : Timed out while waiting for systemd to remove kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice" Mar 17 16:05:23 crc kubenswrapper[4767]: I0317 16:05:23.745191 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-h4pcz" Mar 17 16:05:23 crc kubenswrapper[4767]: I0317 16:05:23.840477 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-h4pcz"] Mar 17 16:05:23 crc kubenswrapper[4767]: I0317 16:05:23.858284 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-h4pcz"] Mar 17 16:05:24 crc kubenswrapper[4767]: I0317 16:05:24.881160 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9kh2v" podUID="65a045e9-219c-4432-9d80-09a944c85013" containerName="registry-server" probeResult="failure" output=< Mar 17 16:05:24 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:05:24 crc kubenswrapper[4767]: > Mar 17 16:05:25 crc kubenswrapper[4767]: I0317 16:05:25.371387 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f4160a8-203a-41e6-a90f-f0800ae776b1" path="/var/lib/kubelet/pods/2f4160a8-203a-41e6-a90f-f0800ae776b1/volumes" Mar 17 16:05:26 crc kubenswrapper[4767]: I0317 16:05:26.966930 4767 generic.go:334] "Generic (PLEG): container finished" podID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerID="786f37d75313e4409a3b52ab9c2aef5ca0cb3e9ad79f97230e225bdf1b660666" exitCode=137 Mar 17 16:05:26 crc kubenswrapper[4767]: I0317 16:05:26.967519 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08838a10-c199-4dc9-bae2-93415ef39ca6","Type":"ContainerDied","Data":"786f37d75313e4409a3b52ab9c2aef5ca0cb3e9ad79f97230e225bdf1b660666"} Mar 17 16:05:27 crc kubenswrapper[4767]: E0317 16:05:27.533864 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified" Mar 17 16:05:27 crc kubenswrapper[4767]: E0317 16:05:27.534627 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ggltx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-dsn5f_openstack(dba94162-c0cc-46e7-a4ca-45836a003a08): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:05:27 crc kubenswrapper[4767]: E0317 16:05:27.535918 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-dsn5f" podUID="dba94162-c0cc-46e7-a4ca-45836a003a08" Mar 17 16:05:28 crc kubenswrapper[4767]: E0317 16:05:28.013447 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-dsn5f" podUID="dba94162-c0cc-46e7-a4ca-45836a003a08" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.560389 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.567566 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.705282 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-combined-ca-bundle\") pod \"302b036a-e22c-4c7d-a924-aa2aa150f81a\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.705407 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-sg-core-conf-yaml\") pod \"08838a10-c199-4dc9-bae2-93415ef39ca6\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.705480 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-run-httpd\") pod \"08838a10-c199-4dc9-bae2-93415ef39ca6\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.705518 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-config-data\") pod \"302b036a-e22c-4c7d-a924-aa2aa150f81a\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.705572 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-combined-ca-bundle\") pod \"08838a10-c199-4dc9-bae2-93415ef39ca6\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.705609 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-scripts\") pod \"302b036a-e22c-4c7d-a924-aa2aa150f81a\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.705770 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-logs\") pod \"302b036a-e22c-4c7d-a924-aa2aa150f81a\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.705808 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-config-data\") pod \"08838a10-c199-4dc9-bae2-93415ef39ca6\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.705843 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-public-tls-certs\") pod \"302b036a-e22c-4c7d-a924-aa2aa150f81a\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.705937 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9mp9\" (UniqueName: \"kubernetes.io/projected/302b036a-e22c-4c7d-a924-aa2aa150f81a-kube-api-access-z9mp9\") pod \"302b036a-e22c-4c7d-a924-aa2aa150f81a\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.709297 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"302b036a-e22c-4c7d-a924-aa2aa150f81a\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.709394 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvc8d\" (UniqueName: \"kubernetes.io/projected/08838a10-c199-4dc9-bae2-93415ef39ca6-kube-api-access-bvc8d\") pod \"08838a10-c199-4dc9-bae2-93415ef39ca6\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.709448 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-log-httpd\") pod \"08838a10-c199-4dc9-bae2-93415ef39ca6\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.709545 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-httpd-run\") pod \"302b036a-e22c-4c7d-a924-aa2aa150f81a\" (UID: \"302b036a-e22c-4c7d-a924-aa2aa150f81a\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.709579 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-scripts\") pod \"08838a10-c199-4dc9-bae2-93415ef39ca6\" (UID: \"08838a10-c199-4dc9-bae2-93415ef39ca6\") " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.712241 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-logs" (OuterVolumeSpecName: "logs") pod "302b036a-e22c-4c7d-a924-aa2aa150f81a" (UID: "302b036a-e22c-4c7d-a924-aa2aa150f81a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.714512 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-scripts" (OuterVolumeSpecName: "scripts") pod "08838a10-c199-4dc9-bae2-93415ef39ca6" (UID: "08838a10-c199-4dc9-bae2-93415ef39ca6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.717717 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "302b036a-e22c-4c7d-a924-aa2aa150f81a" (UID: "302b036a-e22c-4c7d-a924-aa2aa150f81a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.719628 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "08838a10-c199-4dc9-bae2-93415ef39ca6" (UID: "08838a10-c199-4dc9-bae2-93415ef39ca6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.720463 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "08838a10-c199-4dc9-bae2-93415ef39ca6" (UID: "08838a10-c199-4dc9-bae2-93415ef39ca6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.722608 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/302b036a-e22c-4c7d-a924-aa2aa150f81a-kube-api-access-z9mp9" (OuterVolumeSpecName: "kube-api-access-z9mp9") pod "302b036a-e22c-4c7d-a924-aa2aa150f81a" (UID: "302b036a-e22c-4c7d-a924-aa2aa150f81a"). InnerVolumeSpecName "kube-api-access-z9mp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.739416 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-scripts" (OuterVolumeSpecName: "scripts") pod "302b036a-e22c-4c7d-a924-aa2aa150f81a" (UID: "302b036a-e22c-4c7d-a924-aa2aa150f81a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.739707 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08838a10-c199-4dc9-bae2-93415ef39ca6-kube-api-access-bvc8d" (OuterVolumeSpecName: "kube-api-access-bvc8d") pod "08838a10-c199-4dc9-bae2-93415ef39ca6" (UID: "08838a10-c199-4dc9-bae2-93415ef39ca6"). InnerVolumeSpecName "kube-api-access-bvc8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.794922 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b" (OuterVolumeSpecName: "glance") pod "302b036a-e22c-4c7d-a924-aa2aa150f81a" (UID: "302b036a-e22c-4c7d-a924-aa2aa150f81a"). InnerVolumeSpecName "pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.812427 4767 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.812474 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.812485 4767 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.812497 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.812510 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302b036a-e22c-4c7d-a924-aa2aa150f81a-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.812522 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9mp9\" (UniqueName: \"kubernetes.io/projected/302b036a-e22c-4c7d-a924-aa2aa150f81a-kube-api-access-z9mp9\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.812585 4767 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") on node \"crc\" " Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.812604 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvc8d\" (UniqueName: \"kubernetes.io/projected/08838a10-c199-4dc9-bae2-93415ef39ca6-kube-api-access-bvc8d\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.812617 4767 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08838a10-c199-4dc9-bae2-93415ef39ca6-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.842522 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "302b036a-e22c-4c7d-a924-aa2aa150f81a" (UID: "302b036a-e22c-4c7d-a924-aa2aa150f81a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.924154 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.942910 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "08838a10-c199-4dc9-bae2-93415ef39ca6" (UID: "08838a10-c199-4dc9-bae2-93415ef39ca6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.966842 4767 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 17 16:05:28 crc kubenswrapper[4767]: I0317 16:05:28.967085 4767 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b") on node "crc" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.046901 4767 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.046946 4767 reconciler_common.go:293] "Volume detached for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.057534 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08838a10-c199-4dc9-bae2-93415ef39ca6" (UID: "08838a10-c199-4dc9-bae2-93415ef39ca6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.099121 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "302b036a-e22c-4c7d-a924-aa2aa150f81a" (UID: "302b036a-e22c-4c7d-a924-aa2aa150f81a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.115094 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"302b036a-e22c-4c7d-a924-aa2aa150f81a","Type":"ContainerDied","Data":"60742eebfd988bd712d7960f07123e36bf490d29d5121dd9af00263459595593"} Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.115203 4767 scope.go:117] "RemoveContainer" containerID="02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.115432 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.132851 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-config-data" (OuterVolumeSpecName: "config-data") pod "302b036a-e22c-4c7d-a924-aa2aa150f81a" (UID: "302b036a-e22c-4c7d-a924-aa2aa150f81a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.155938 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.155997 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.156013 4767 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/302b036a-e22c-4c7d-a924-aa2aa150f81a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.162710 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08838a10-c199-4dc9-bae2-93415ef39ca6","Type":"ContainerDied","Data":"616cc83e4a78acb06cef32fda4bd694ffe10a73cb05346c7f0213a5728220c5c"} Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.162855 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.167838 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-config-data" (OuterVolumeSpecName: "config-data") pod "08838a10-c199-4dc9-bae2-93415ef39ca6" (UID: "08838a10-c199-4dc9-bae2-93415ef39ca6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.258029 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08838a10-c199-4dc9-bae2-93415ef39ca6-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.457465 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.478247 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.509825 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.538518 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.553752 4767 scope.go:117] "RemoveContainer" containerID="f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.573235 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574002 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerName="heat-api" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574030 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerName="heat-api" Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574056 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="proxy-httpd" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574062 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="proxy-httpd" Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574077 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="ceilometer-central-agent" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574084 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="ceilometer-central-agent" Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574108 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="302b036a-e22c-4c7d-a924-aa2aa150f81a" containerName="glance-httpd" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574115 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="302b036a-e22c-4c7d-a924-aa2aa150f81a" containerName="glance-httpd" Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574137 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2610e145-1410-4330-bcca-808faf035214" containerName="heat-engine" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574143 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2610e145-1410-4330-bcca-808faf035214" containerName="heat-engine" Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574179 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="ceilometer-notification-agent" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574185 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="ceilometer-notification-agent" Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574192 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="sg-core" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574198 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="sg-core" Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574210 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" containerName="heat-cfnapi" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574215 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" containerName="heat-cfnapi" Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574228 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" containerName="heat-cfnapi" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574234 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" containerName="heat-cfnapi" Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574245 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerName="heat-api" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574252 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerName="heat-api" Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574263 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="302b036a-e22c-4c7d-a924-aa2aa150f81a" containerName="glance-log" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574271 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="302b036a-e22c-4c7d-a924-aa2aa150f81a" containerName="glance-log" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574574 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerName="heat-api" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574595 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerName="heat-api" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574604 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="302b036a-e22c-4c7d-a924-aa2aa150f81a" containerName="glance-log" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574617 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="302b036a-e22c-4c7d-a924-aa2aa150f81a" containerName="glance-httpd" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574625 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="proxy-httpd" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574639 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="ceilometer-central-agent" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574649 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="2610e145-1410-4330-bcca-808faf035214" containerName="heat-engine" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574659 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" containerName="heat-cfnapi" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574674 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="ceilometer-notification-agent" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574683 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" containerName="sg-core" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574695 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9c5ebc5-6ab9-4d98-9a82-0f3c0752906c" containerName="heat-cfnapi" Mar 17 16:05:29 crc kubenswrapper[4767]: E0317 16:05:29.574921 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerName="heat-api" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.574930 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerName="heat-api" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.575183 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="33f239ab-85ed-4803-a0db-c31dc5301d28" containerName="heat-api" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.576427 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.585482 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.587022 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.600371 4767 scope.go:117] "RemoveContainer" containerID="786f37d75313e4409a3b52ab9c2aef5ca0cb3e9ad79f97230e225bdf1b660666" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.632357 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.640028 4767 scope.go:117] "RemoveContainer" containerID="19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.655059 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.669542 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.669705 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.674892 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.675056 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.684873 4767 scope.go:117] "RemoveContainer" containerID="0bd1b41c99c5fee3341231a56e77f3f105618713dca3db21ef95535978c355eb" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.708623 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.709020 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.709432 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-config-data\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.709615 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-scripts\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.709769 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7mxb\" (UniqueName: \"kubernetes.io/projected/824d9a59-5de1-4ead-9bc4-09de9253677e-kube-api-access-q7mxb\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.709807 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/824d9a59-5de1-4ead-9bc4-09de9253677e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.709957 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/824d9a59-5de1-4ead-9bc4-09de9253677e-logs\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.714152 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.731511 4767 scope.go:117] "RemoveContainer" containerID="390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.817336 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-log-httpd\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.817469 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.817513 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-config-data\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.817585 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.817614 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.818013 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.818090 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-config-data\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.818200 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-scripts\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.818337 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-scripts\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.818512 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.818545 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7mxb\" (UniqueName: \"kubernetes.io/projected/824d9a59-5de1-4ead-9bc4-09de9253677e-kube-api-access-q7mxb\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.818582 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/824d9a59-5de1-4ead-9bc4-09de9253677e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.818698 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-run-httpd\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.818738 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75dg6\" (UniqueName: \"kubernetes.io/projected/65f3e07f-47cf-4973-82e9-353692ec4a0b-kube-api-access-75dg6\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.818804 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/824d9a59-5de1-4ead-9bc4-09de9253677e-logs\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.819283 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/824d9a59-5de1-4ead-9bc4-09de9253677e-logs\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.821259 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/824d9a59-5de1-4ead-9bc4-09de9253677e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.823627 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-scripts\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.824520 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.825433 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-config-data\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.825620 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824d9a59-5de1-4ead-9bc4-09de9253677e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.841228 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.841293 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/683557e62de6f02370abe27ab62c9a425f8a422dda40ecee534d661837e8035d/globalmount\"" pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.846964 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7mxb\" (UniqueName: \"kubernetes.io/projected/824d9a59-5de1-4ead-9bc4-09de9253677e-kube-api-access-q7mxb\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.905084 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6b71e918-ff4a-47ae-9348-9663d9f8d58b\") pod \"glance-default-external-api-0\" (UID: \"824d9a59-5de1-4ead-9bc4-09de9253677e\") " pod="openstack/glance-default-external-api-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.921603 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-scripts\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.921751 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.921805 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-run-httpd\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.921828 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75dg6\" (UniqueName: \"kubernetes.io/projected/65f3e07f-47cf-4973-82e9-353692ec4a0b-kube-api-access-75dg6\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.921858 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-log-httpd\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.921887 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-config-data\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.921923 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.923120 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-run-httpd\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.923701 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-log-httpd\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.928261 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-config-data\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.929191 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.929214 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-scripts\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.929339 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:29 crc kubenswrapper[4767]: I0317 16:05:29.945258 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75dg6\" (UniqueName: \"kubernetes.io/projected/65f3e07f-47cf-4973-82e9-353692ec4a0b-kube-api-access-75dg6\") pod \"ceilometer-0\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " pod="openstack/ceilometer-0" Mar 17 16:05:30 crc kubenswrapper[4767]: I0317 16:05:30.014607 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:05:30 crc kubenswrapper[4767]: I0317 16:05:30.201121 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 16:05:30 crc kubenswrapper[4767]: I0317 16:05:30.634997 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:05:30 crc kubenswrapper[4767]: I0317 16:05:30.956991 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 16:05:31 crc kubenswrapper[4767]: I0317 16:05:31.469588 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:05:31 crc kubenswrapper[4767]: E0317 16:05:31.470038 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:05:31 crc kubenswrapper[4767]: I0317 16:05:31.518828 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08838a10-c199-4dc9-bae2-93415ef39ca6" path="/var/lib/kubelet/pods/08838a10-c199-4dc9-bae2-93415ef39ca6/volumes" Mar 17 16:05:31 crc kubenswrapper[4767]: I0317 16:05:31.520006 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="302b036a-e22c-4c7d-a924-aa2aa150f81a" path="/var/lib/kubelet/pods/302b036a-e22c-4c7d-a924-aa2aa150f81a/volumes" Mar 17 16:05:31 crc kubenswrapper[4767]: I0317 16:05:31.527726 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65f3e07f-47cf-4973-82e9-353692ec4a0b","Type":"ContainerStarted","Data":"7088ba7b37e5f96402f31a68ef2fbaea4383aa9aee7ec624b38e6cec0dd49c64"} Mar 17 16:05:31 crc kubenswrapper[4767]: I0317 16:05:31.531729 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"824d9a59-5de1-4ead-9bc4-09de9253677e","Type":"ContainerStarted","Data":"a1c9ccbf57b6cef27f49881c90db27d034e4c210e1e1b09189b984891c5c7365"} Mar 17 16:05:32 crc kubenswrapper[4767]: I0317 16:05:32.577869 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65f3e07f-47cf-4973-82e9-353692ec4a0b","Type":"ContainerStarted","Data":"02333d1ad02cc38464ce2d996650bc5058ce4ce600c1ffaeb12bc4deee18a0c3"} Mar 17 16:05:32 crc kubenswrapper[4767]: I0317 16:05:32.581969 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"824d9a59-5de1-4ead-9bc4-09de9253677e","Type":"ContainerStarted","Data":"f5771bc836db9d74cc988becfb5aa69a3d8640a24d4673bca7bcadcc1fe785c1"} Mar 17 16:05:33 crc kubenswrapper[4767]: I0317 16:05:33.601408 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65f3e07f-47cf-4973-82e9-353692ec4a0b","Type":"ContainerStarted","Data":"ed6d5c7f2bdbef020415709a1c9d8bba43d9bf3780d06686c9f18e737513854c"} Mar 17 16:05:33 crc kubenswrapper[4767]: I0317 16:05:33.603963 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"824d9a59-5de1-4ead-9bc4-09de9253677e","Type":"ContainerStarted","Data":"9f39311220e9f24715849bba797813769d7dec38f8c7bde4e7023aba66092009"} Mar 17 16:05:33 crc kubenswrapper[4767]: I0317 16:05:33.637447 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.637410384 podStartE2EDuration="4.637410384s" podCreationTimestamp="2026-03-17 16:05:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:05:33.629352294 +0000 UTC m=+1725.042668341" watchObservedRunningTime="2026-03-17 16:05:33.637410384 +0000 UTC m=+1725.050726431" Mar 17 16:05:33 crc kubenswrapper[4767]: I0317 16:05:33.885565 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:05:33 crc kubenswrapper[4767]: I0317 16:05:33.951877 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:05:34 crc kubenswrapper[4767]: I0317 16:05:34.135313 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9kh2v"] Mar 17 16:05:34 crc kubenswrapper[4767]: I0317 16:05:34.406831 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:05:34 crc kubenswrapper[4767]: E0317 16:05:34.545904 4767 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/9e2ce6fdc24ee9b83b1abe217e39cf7c6a382ee0595bf6e63593e17f6f4b8dff/diff" to get inode usage: stat /var/lib/containers/storage/overlay/9e2ce6fdc24ee9b83b1abe217e39cf7c6a382ee0595bf6e63593e17f6f4b8dff/diff: no such file or directory, extraDiskErr: Mar 17 16:05:34 crc kubenswrapper[4767]: I0317 16:05:34.620596 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65f3e07f-47cf-4973-82e9-353692ec4a0b","Type":"ContainerStarted","Data":"61fa65eb2e6a84d997f3520cb68e39294a7eb109dfff6c4bdd37368a839d2767"} Mar 17 16:05:34 crc kubenswrapper[4767]: E0317 16:05:34.942127 4767 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/9573bce0624e95bd0edf28f5ce515cbc6ea9fa3588b30d21cbd1c5cee5de7637/diff" to get inode usage: stat /var/lib/containers/storage/overlay/9573bce0624e95bd0edf28f5ce515cbc6ea9fa3588b30d21cbd1c5cee5de7637/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_heat-engine-66ff489c4c-47nsf_2610e145-1410-4330-bcca-808faf035214/heat-engine/0.log" to get inode usage: stat /var/log/pods/openstack_heat-engine-66ff489c4c-47nsf_2610e145-1410-4330-bcca-808faf035214/heat-engine/0.log: no such file or directory Mar 17 16:05:35 crc kubenswrapper[4767]: I0317 16:05:35.882569 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9kh2v" podUID="65a045e9-219c-4432-9d80-09a944c85013" containerName="registry-server" containerID="cri-o://70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41" gracePeriod=2 Mar 17 16:05:36 crc kubenswrapper[4767]: E0317 16:05:36.077359 4767 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/64bc6523e53f163ba134254b038c450f68a2d9eed532035daa8eabd3a759be45/diff" to get inode usage: stat /var/lib/containers/storage/overlay/64bc6523e53f163ba134254b038c450f68a2d9eed532035daa8eabd3a759be45/diff: no such file or directory, extraDiskErr: Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.671583 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.766777 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-utilities\") pod \"65a045e9-219c-4432-9d80-09a944c85013\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.766868 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-catalog-content\") pod \"65a045e9-219c-4432-9d80-09a944c85013\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.766992 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxjdt\" (UniqueName: \"kubernetes.io/projected/65a045e9-219c-4432-9d80-09a944c85013-kube-api-access-nxjdt\") pod \"65a045e9-219c-4432-9d80-09a944c85013\" (UID: \"65a045e9-219c-4432-9d80-09a944c85013\") " Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.767545 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-utilities" (OuterVolumeSpecName: "utilities") pod "65a045e9-219c-4432-9d80-09a944c85013" (UID: "65a045e9-219c-4432-9d80-09a944c85013"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.767894 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.775547 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65a045e9-219c-4432-9d80-09a944c85013-kube-api-access-nxjdt" (OuterVolumeSpecName: "kube-api-access-nxjdt") pod "65a045e9-219c-4432-9d80-09a944c85013" (UID: "65a045e9-219c-4432-9d80-09a944c85013"). InnerVolumeSpecName "kube-api-access-nxjdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.843255 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65a045e9-219c-4432-9d80-09a944c85013" (UID: "65a045e9-219c-4432-9d80-09a944c85013"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.870854 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65a045e9-219c-4432-9d80-09a944c85013-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.870906 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxjdt\" (UniqueName: \"kubernetes.io/projected/65a045e9-219c-4432-9d80-09a944c85013-kube-api-access-nxjdt\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.898571 4767 generic.go:334] "Generic (PLEG): container finished" podID="65a045e9-219c-4432-9d80-09a944c85013" containerID="70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41" exitCode=0 Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.898657 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9kh2v" Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.898685 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kh2v" event={"ID":"65a045e9-219c-4432-9d80-09a944c85013","Type":"ContainerDied","Data":"70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41"} Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.898731 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9kh2v" event={"ID":"65a045e9-219c-4432-9d80-09a944c85013","Type":"ContainerDied","Data":"1d5e03d5f46a6b457764cbf327336b3881a4b232d151d5ab3221aaaacf8fdfec"} Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.898750 4767 scope.go:117] "RemoveContainer" containerID="70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41" Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.933303 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65f3e07f-47cf-4973-82e9-353692ec4a0b","Type":"ContainerStarted","Data":"bffa8f087ec5113fb13fa1cd880317f4f67679bd0724b257a717a09ba28c4cb2"} Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.933534 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="ceilometer-central-agent" containerID="cri-o://02333d1ad02cc38464ce2d996650bc5058ce4ce600c1ffaeb12bc4deee18a0c3" gracePeriod=30 Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.933669 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.933716 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="proxy-httpd" containerID="cri-o://bffa8f087ec5113fb13fa1cd880317f4f67679bd0724b257a717a09ba28c4cb2" gracePeriod=30 Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.933785 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="ceilometer-notification-agent" containerID="cri-o://ed6d5c7f2bdbef020415709a1c9d8bba43d9bf3780d06686c9f18e737513854c" gracePeriod=30 Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.933914 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="sg-core" containerID="cri-o://61fa65eb2e6a84d997f3520cb68e39294a7eb109dfff6c4bdd37368a839d2767" gracePeriod=30 Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.969354 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9kh2v"] Mar 17 16:05:36 crc kubenswrapper[4767]: I0317 16:05:36.997109 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9kh2v"] Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.005074 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.185245509 podStartE2EDuration="8.005028884s" podCreationTimestamp="2026-03-17 16:05:29 +0000 UTC" firstStartedPulling="2026-03-17 16:05:30.639493509 +0000 UTC m=+1722.052809556" lastFinishedPulling="2026-03-17 16:05:36.459276884 +0000 UTC m=+1727.872592931" observedRunningTime="2026-03-17 16:05:36.987515143 +0000 UTC m=+1728.400831210" watchObservedRunningTime="2026-03-17 16:05:37.005028884 +0000 UTC m=+1728.418344931" Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.013549 4767 scope.go:117] "RemoveContainer" containerID="f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004" Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.060696 4767 scope.go:117] "RemoveContainer" containerID="3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191" Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.211821 4767 scope.go:117] "RemoveContainer" containerID="70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41" Mar 17 16:05:37 crc kubenswrapper[4767]: E0317 16:05:37.212403 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41\": container with ID starting with 70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41 not found: ID does not exist" containerID="70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41" Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.212463 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41"} err="failed to get container status \"70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41\": rpc error: code = NotFound desc = could not find container \"70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41\": container with ID starting with 70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41 not found: ID does not exist" Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.212506 4767 scope.go:117] "RemoveContainer" containerID="f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004" Mar 17 16:05:37 crc kubenswrapper[4767]: E0317 16:05:37.212870 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004\": container with ID starting with f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004 not found: ID does not exist" containerID="f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004" Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.212902 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004"} err="failed to get container status \"f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004\": rpc error: code = NotFound desc = could not find container \"f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004\": container with ID starting with f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004 not found: ID does not exist" Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.212920 4767 scope.go:117] "RemoveContainer" containerID="3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191" Mar 17 16:05:37 crc kubenswrapper[4767]: E0317 16:05:37.213328 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191\": container with ID starting with 3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191 not found: ID does not exist" containerID="3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191" Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.213366 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191"} err="failed to get container status \"3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191\": rpc error: code = NotFound desc = could not find container \"3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191\": container with ID starting with 3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191 not found: ID does not exist" Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.371672 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65a045e9-219c-4432-9d80-09a944c85013" path="/var/lib/kubelet/pods/65a045e9-219c-4432-9d80-09a944c85013/volumes" Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.951227 4767 generic.go:334] "Generic (PLEG): container finished" podID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerID="61fa65eb2e6a84d997f3520cb68e39294a7eb109dfff6c4bdd37368a839d2767" exitCode=2 Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.951569 4767 generic.go:334] "Generic (PLEG): container finished" podID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerID="ed6d5c7f2bdbef020415709a1c9d8bba43d9bf3780d06686c9f18e737513854c" exitCode=0 Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.951310 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65f3e07f-47cf-4973-82e9-353692ec4a0b","Type":"ContainerDied","Data":"61fa65eb2e6a84d997f3520cb68e39294a7eb109dfff6c4bdd37368a839d2767"} Mar 17 16:05:37 crc kubenswrapper[4767]: I0317 16:05:37.951615 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65f3e07f-47cf-4973-82e9-353692ec4a0b","Type":"ContainerDied","Data":"ed6d5c7f2bdbef020415709a1c9d8bba43d9bf3780d06686c9f18e737513854c"} Mar 17 16:05:38 crc kubenswrapper[4767]: I0317 16:05:38.967246 4767 generic.go:334] "Generic (PLEG): container finished" podID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerID="02333d1ad02cc38464ce2d996650bc5058ce4ce600c1ffaeb12bc4deee18a0c3" exitCode=0 Mar 17 16:05:38 crc kubenswrapper[4767]: I0317 16:05:38.967341 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65f3e07f-47cf-4973-82e9-353692ec4a0b","Type":"ContainerDied","Data":"02333d1ad02cc38464ce2d996650bc5058ce4ce600c1ffaeb12bc4deee18a0c3"} Mar 17 16:05:40 crc kubenswrapper[4767]: E0317 16:05:40.052204 4767 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/7ea0e92a09361ea828000f67b82120f837e223027e2a83a3ce7770ebf372fbc2/diff" to get inode usage: stat /var/lib/containers/storage/overlay/7ea0e92a09361ea828000f67b82120f837e223027e2a83a3ce7770ebf372fbc2/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_08838a10-c199-4dc9-bae2-93415ef39ca6/ceilometer-central-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_08838a10-c199-4dc9-bae2-93415ef39ca6/ceilometer-central-agent/0.log: no such file or directory Mar 17 16:05:40 crc kubenswrapper[4767]: I0317 16:05:40.201391 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 17 16:05:40 crc kubenswrapper[4767]: I0317 16:05:40.201475 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 17 16:05:40 crc kubenswrapper[4767]: I0317 16:05:40.248438 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 17 16:05:40 crc kubenswrapper[4767]: I0317 16:05:40.263570 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 17 16:05:40 crc kubenswrapper[4767]: I0317 16:05:40.994335 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 17 16:05:40 crc kubenswrapper[4767]: I0317 16:05:40.994391 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 17 16:05:42 crc kubenswrapper[4767]: E0317 16:05:42.100359 4767 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/4988669516fa6e37a9a801ab163f54436e357ece73e84e59e01353ac9fe3f7aa/diff" to get inode usage: stat /var/lib/containers/storage/overlay/4988669516fa6e37a9a801ab163f54436e357ece73e84e59e01353ac9fe3f7aa/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_08838a10-c199-4dc9-bae2-93415ef39ca6/ceilometer-notification-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_08838a10-c199-4dc9-bae2-93415ef39ca6/ceilometer-notification-agent/0.log: no such file or directory Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.155931 4767 generic.go:334] "Generic (PLEG): container finished" podID="00a982e7-017e-4f3c-a8aa-21eab359cad4" containerID="cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db" exitCode=137 Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.156242 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-658d967468-8hprg" event={"ID":"00a982e7-017e-4f3c-a8aa-21eab359cad4","Type":"ContainerDied","Data":"cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db"} Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.654219 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:05:42 crc kubenswrapper[4767]: W0317 16:05:42.737871 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-1d5e03d5f46a6b457764cbf327336b3881a4b232d151d5ab3221aaaacf8fdfec": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-1d5e03d5f46a6b457764cbf327336b3881a4b232d151d5ab3221aaaacf8fdfec: no such file or directory Mar 17 16:05:42 crc kubenswrapper[4767]: W0317 16:05:42.737921 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-conmon-3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-conmon-3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191.scope: no such file or directory Mar 17 16:05:42 crc kubenswrapper[4767]: W0317 16:05:42.737947 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-3cebc213cd1904e6e80b041a78a2b8408118f4cf1d1e37062c32c6dce5e41191.scope: no such file or directory Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.742933 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-combined-ca-bundle\") pod \"00a982e7-017e-4f3c-a8aa-21eab359cad4\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.743445 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data-custom\") pod \"00a982e7-017e-4f3c-a8aa-21eab359cad4\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.743581 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9twc\" (UniqueName: \"kubernetes.io/projected/00a982e7-017e-4f3c-a8aa-21eab359cad4-kube-api-access-r9twc\") pod \"00a982e7-017e-4f3c-a8aa-21eab359cad4\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.743741 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data\") pod \"00a982e7-017e-4f3c-a8aa-21eab359cad4\" (UID: \"00a982e7-017e-4f3c-a8aa-21eab359cad4\") " Mar 17 16:05:42 crc kubenswrapper[4767]: W0317 16:05:42.752927 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-786f37d75313e4409a3b52ab9c2aef5ca0cb3e9ad79f97230e225bdf1b660666.scope WatchSource:0}: Error finding container 786f37d75313e4409a3b52ab9c2aef5ca0cb3e9ad79f97230e225bdf1b660666: Status 404 returned error can't find the container with id 786f37d75313e4409a3b52ab9c2aef5ca0cb3e9ad79f97230e225bdf1b660666 Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.762261 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "00a982e7-017e-4f3c-a8aa-21eab359cad4" (UID: "00a982e7-017e-4f3c-a8aa-21eab359cad4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.764054 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a982e7-017e-4f3c-a8aa-21eab359cad4-kube-api-access-r9twc" (OuterVolumeSpecName: "kube-api-access-r9twc") pod "00a982e7-017e-4f3c-a8aa-21eab359cad4" (UID: "00a982e7-017e-4f3c-a8aa-21eab359cad4"). InnerVolumeSpecName "kube-api-access-r9twc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:05:42 crc kubenswrapper[4767]: W0317 16:05:42.790165 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-conmon-f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-conmon-f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004.scope: no such file or directory Mar 17 16:05:42 crc kubenswrapper[4767]: W0317 16:05:42.790366 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-f39e3055cdca47d0e4c484cc3c838cd9f03462116e9eb453bb43c16d3f1aa004.scope: no such file or directory Mar 17 16:05:42 crc kubenswrapper[4767]: W0317 16:05:42.790987 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice/crio-conmon-6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice/crio-conmon-6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d.scope: no such file or directory Mar 17 16:05:42 crc kubenswrapper[4767]: W0317 16:05:42.791048 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-conmon-70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-conmon-70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41.scope: no such file or directory Mar 17 16:05:42 crc kubenswrapper[4767]: W0317 16:05:42.791068 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice/crio-6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice/crio-6f3cf9fcb7cca0f36250dbadaaf8d34192a85973cb539a71e6cb06c72115869d.scope: no such file or directory Mar 17 16:05:42 crc kubenswrapper[4767]: W0317 16:05:42.791218 4767 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a045e9_219c_4432_9d80_09a944c85013.slice/crio-70889e74c381c578976aa46f22a9f3b353204baed402d10ce961374618caca41.scope: no such file or directory Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.830469 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00a982e7-017e-4f3c-a8aa-21eab359cad4" (UID: "00a982e7-017e-4f3c-a8aa-21eab359cad4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.857282 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.857354 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9twc\" (UniqueName: \"kubernetes.io/projected/00a982e7-017e-4f3c-a8aa-21eab359cad4-kube-api-access-r9twc\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.857368 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.870466 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data" (OuterVolumeSpecName: "config-data") pod "00a982e7-017e-4f3c-a8aa-21eab359cad4" (UID: "00a982e7-017e-4f3c-a8aa-21eab359cad4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:42 crc kubenswrapper[4767]: E0317 16:05:42.905437 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-conmon-71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1137107_c196_4307_b2c4_90a6e9a0aab4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-conmon-390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-e4f1fc5d421dc2bd9c32944f148850e3c7906c603d6614c995b39964a2b95272\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-conmon-02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9a6a8fc_9770_4e16_8f14_371b008d3742.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-conmon-e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fe2d6_1df4_4412_a107_a12bcb38e91f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-conmon-cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-conmon-19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24e92570_fb4d_4fa2_ae4f_cc11da8f764a.slice\": RecentStats: unable to find data in memory cache], [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice/crio-19efcb8894fdf242e184e7df57c2455f700702871ff1c0f5f0d9be62fa6860f2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24e92570_fb4d_4fa2_ae4f_cc11da8f764a.slice/crio-9c389228536150f157fedb29d96e58bb0fcdd2211a266a61a40f255ba8e1aa01\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-616cc83e4a78acb06cef32fda4bd694ffe10a73cb05346c7f0213a5728220c5c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-0bd1b41c99c5fee3341231a56e77f3f105618713dca3db21ef95535978c355eb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c5ebc5_6ab9_4d98_9a82_0f3c0752906c.slice/crio-ee10d09d620425ffc6e31e2cff117403df18c283fb631367bbb5d33b7d5bc463\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-conmon-234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fe2d6_1df4_4412_a107_a12bcb38e91f.slice/crio-9047fd2850cafa2664014a183b86d91d0b386c26afaebcf2407a7e02440b02bb\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-60310b999086a1beb533258ea23e383a8571ce64431b108a386658cc08d1b5b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-a9f3df5fc86345edfe138aefdfbaeeb838ca491f039c7d110fb7e6384f38eb9e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice/crio-87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c5ebc5_6ab9_4d98_9a82_0f3c0752906c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fe2d6_1df4_4412_a107_a12bcb38e91f.slice/crio-conmon-72df4826ccc3d3f9a8847830042fbbc37df929e953b73a207af394e476ca5a43.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-60742eebfd988bd712d7960f07123e36bf490d29d5121dd9af00263459595593\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice/crio-10689c0e51f1e6d4f4030b8fd5f677031eaedf4ed42a2d43f5141213dab52fc7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-conmon-f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice/crio-89469eed3c5666995f50f5309cf2b253b49b11c04e7890fe945935f52393344d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice/crio-conmon-87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-conmon-0bd1b41c99c5fee3341231a56e77f3f105618713dca3db21ef95535978c355eb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:05:42 crc kubenswrapper[4767]: E0317 16:05:42.905841 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fe2d6_1df4_4412_a107_a12bcb38e91f.slice/crio-9047fd2850cafa2664014a183b86d91d0b386c26afaebcf2407a7e02440b02bb\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice/crio-10689c0e51f1e6d4f4030b8fd5f677031eaedf4ed42a2d43f5141213dab52fc7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-conmon-e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-60310b999086a1beb533258ea23e383a8571ce64431b108a386658cc08d1b5b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-0bd1b41c99c5fee3341231a56e77f3f105618713dca3db21ef95535978c355eb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice/crio-19efcb8894fdf242e184e7df57c2455f700702871ff1c0f5f0d9be62fa6860f2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice/crio-89469eed3c5666995f50f5309cf2b253b49b11c04e7890fe945935f52393344d\": RecentStats: unable to find data in memory cache], [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-60742eebfd988bd712d7960f07123e36bf490d29d5121dd9af00263459595593\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-conmon-234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-e4f1fc5d421dc2bd9c32944f148850e3c7906c603d6614c995b39964a2b95272\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24e92570_fb4d_4fa2_ae4f_cc11da8f764a.slice/crio-9c389228536150f157fedb29d96e58bb0fcdd2211a266a61a40f255ba8e1aa01\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-conmon-19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c5ebc5_6ab9_4d98_9a82_0f3c0752906c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c5ebc5_6ab9_4d98_9a82_0f3c0752906c.slice/crio-ee10d09d620425ffc6e31e2cff117403df18c283fb631367bbb5d33b7d5bc463\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9a6a8fc_9770_4e16_8f14_371b008d3742.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-conmon-0bd1b41c99c5fee3341231a56e77f3f105618713dca3db21ef95535978c355eb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1137107_c196_4307_b2c4_90a6e9a0aab4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-a9f3df5fc86345edfe138aefdfbaeeb838ca491f039c7d110fb7e6384f38eb9e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9a6a8fc_9770_4e16_8f14_371b008d3742.slice/crio-898ed8832f6b941b5bb20f9c1244dc76126c914f33a93cb74a5a34c4407a8018\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fe2d6_1df4_4412_a107_a12bcb38e91f.slice/crio-conmon-72df4826ccc3d3f9a8847830042fbbc37df929e953b73a207af394e476ca5a43.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice/crio-conmon-87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-conmon-cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-616cc83e4a78acb06cef32fda4bd694ffe10a73cb05346c7f0213a5728220c5c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice/crio-87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-conmon-f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-conmon-71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-conmon-02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24e92570_fb4d_4fa2_ae4f_cc11da8f764a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-conmon-390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678.scope\": RecentStats: unable to find data in memory cache]" Mar 17 16:05:42 crc kubenswrapper[4767]: E0317 16:05:42.908065 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24e92570_fb4d_4fa2_ae4f_cc11da8f764a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1137107_c196_4307_b2c4_90a6e9a0aab4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-60742eebfd988bd712d7960f07123e36bf490d29d5121dd9af00263459595593\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-60310b999086a1beb533258ea23e383a8571ce64431b108a386658cc08d1b5b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice/crio-conmon-87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-conmon-19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24e92570_fb4d_4fa2_ae4f_cc11da8f764a.slice/crio-9c389228536150f157fedb29d96e58bb0fcdd2211a266a61a40f255ba8e1aa01\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-conmon-02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice/crio-19efcb8894fdf242e184e7df57c2455f700702871ff1c0f5f0d9be62fa6860f2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-e4f1fc5d421dc2bd9c32944f148850e3c7906c603d6614c995b39964a2b95272\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-a9f3df5fc86345edfe138aefdfbaeeb838ca491f039c7d110fb7e6384f38eb9e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fe2d6_1df4_4412_a107_a12bcb38e91f.slice/crio-conmon-72df4826ccc3d3f9a8847830042fbbc37df929e953b73a207af394e476ca5a43.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c5ebc5_6ab9_4d98_9a82_0f3c0752906c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-conmon-234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-616cc83e4a78acb06cef32fda4bd694ffe10a73cb05346c7f0213a5728220c5c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-conmon-390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c5ebc5_6ab9_4d98_9a82_0f3c0752906c.slice/crio-ee10d09d620425ffc6e31e2cff117403df18c283fb631367bbb5d33b7d5bc463\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice/crio-10689c0e51f1e6d4f4030b8fd5f677031eaedf4ed42a2d43f5141213dab52fc7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-conmon-f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-0bd1b41c99c5fee3341231a56e77f3f105618713dca3db21ef95535978c355eb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9a6a8fc_9770_4e16_8f14_371b008d3742.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fe2d6_1df4_4412_a107_a12bcb38e91f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9a6a8fc_9770_4e16_8f14_371b008d3742.slice/crio-898ed8832f6b941b5bb20f9c1244dc76126c914f33a93cb74a5a34c4407a8018\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-conmon-cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-conmon-71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice/crio-89469eed3c5666995f50f5309cf2b253b49b11c04e7890fe945935f52393344d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fe2d6_1df4_4412_a107_a12bcb38e91f.slice/crio-9047fd2850cafa2664014a183b86d91d0b386c26afaebcf2407a7e02440b02bb\": RecentStats: unable to find data in memory cache]" Mar 17 16:05:42 crc kubenswrapper[4767]: E0317 16:05:42.915140 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice/crio-19efcb8894fdf242e184e7df57c2455f700702871ff1c0f5f0d9be62fa6860f2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-60742eebfd988bd712d7960f07123e36bf490d29d5121dd9af00263459595593\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24e92570_fb4d_4fa2_ae4f_cc11da8f764a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice/crio-10689c0e51f1e6d4f4030b8fd5f677031eaedf4ed42a2d43f5141213dab52fc7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-conmon-e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-a9f3df5fc86345edfe138aefdfbaeeb838ca491f039c7d110fb7e6384f38eb9e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice/crio-87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-conmon-19912de08f78c8d044d1145547ef465f583e2335bcfda3d84a8150c77cc0e09c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-e4f1fc5d421dc2bd9c32944f148850e3c7906c603d6614c995b39964a2b95272\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9a6a8fc_9770_4e16_8f14_371b008d3742.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-conmon-cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fe2d6_1df4_4412_a107_a12bcb38e91f.slice/crio-conmon-72df4826ccc3d3f9a8847830042fbbc37df929e953b73a207af394e476ca5a43.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4160a8_203a_41e6_a90f_f0800ae776b1.slice/crio-89469eed3c5666995f50f5309cf2b253b49b11c04e7890fe945935f52393344d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d5a577_7603_4cfe_a0c9_a899eeaab93d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24e92570_fb4d_4fa2_ae4f_cc11da8f764a.slice/crio-9c389228536150f157fedb29d96e58bb0fcdd2211a266a61a40f255ba8e1aa01\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-0bd1b41c99c5fee3341231a56e77f3f105618713dca3db21ef95535978c355eb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-e3f487c76ee89255a27679c1a25b0bc47d8b68f42ad065cb3d6a811d9a3dd04f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fe2d6_1df4_4412_a107_a12bcb38e91f.slice/crio-9047fd2850cafa2664014a183b86d91d0b386c26afaebcf2407a7e02440b02bb\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610e145_1410_4330_bcca_808faf035214.slice/crio-conmon-87e9183e83a823f468f095037eb2f93f3a5be82dbf05aedb47a3d63b98a2fdb7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9a6a8fc_9770_4e16_8f14_371b008d3742.slice/crio-898ed8832f6b941b5bb20f9c1244dc76126c914f33a93cb74a5a34c4407a8018\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-conmon-390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-conmon-02471fbd139fb89f1c0bbc7a0b18317582503d1a13f83486aaebf7e9cced864e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33f239ab_85ed_4803_a0db_c31dc5301d28.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-616cc83e4a78acb06cef32fda4bd694ffe10a73cb05346c7f0213a5728220c5c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb414e8_1117_4306_85b8_311102e2d8fb.slice/crio-conmon-234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e8a6ca_f669_4eef_8509_f30f7ee31447.slice/crio-conmon-71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c5ebc5_6ab9_4d98_9a82_0f3c0752906c.slice/crio-ee10d09d620425ffc6e31e2cff117403df18c283fb631367bbb5d33b7d5bc463\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65779f05_d74b_49a7_a31a_2de7f9005e59.slice/crio-60310b999086a1beb533258ea23e383a8571ce64431b108a386658cc08d1b5b9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302b036a_e22c_4c7d_a924_aa2aa150f81a.slice/crio-conmon-f5bf570b7f377e989e0fc8995dfdf31eef6fcc750694d8365b0b9c7acd238156.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08838a10_c199_4dc9_bae2_93415ef39ca6.slice/crio-390a4862d4a70887ff20a00f8625cd169801b9f1c823a6fd2247d08adc0a8678.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c5ebc5_6ab9_4d98_9a82_0f3c0752906c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1137107_c196_4307_b2c4_90a6e9a0aab4.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:05:42 crc kubenswrapper[4767]: I0317 16:05:42.961739 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a982e7-017e-4f3c-a8aa-21eab359cad4-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.174575 4767 generic.go:334] "Generic (PLEG): container finished" podID="3caf16d7-044a-4efc-a3ba-c494255df8ff" containerID="b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e" exitCode=137 Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.174668 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-75777db76d-5z4qz" event={"ID":"3caf16d7-044a-4efc-a3ba-c494255df8ff","Type":"ContainerDied","Data":"b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e"} Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.177408 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dsn5f" event={"ID":"dba94162-c0cc-46e7-a4ca-45836a003a08","Type":"ContainerStarted","Data":"19902911191b266a4117332e9664ffbe0b11a0de4db4a9275ffc9802b5c13c9e"} Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.185045 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-658d967468-8hprg" event={"ID":"00a982e7-017e-4f3c-a8aa-21eab359cad4","Type":"ContainerDied","Data":"f646134bf7c3e81d39895e638d392ddf0d379219ae5e27f819f2a8aecd6a8030"} Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.185130 4767 scope.go:117] "RemoveContainer" containerID="cfe588d0cd0541339c5d7d43fa6422eec69da2e4f4b881dccd519c5be75f39db" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.185311 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-658d967468-8hprg" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.214588 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.217350 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-dsn5f" podStartSLOduration=4.054237686 podStartE2EDuration="46.217325085s" podCreationTimestamp="2026-03-17 16:04:57 +0000 UTC" firstStartedPulling="2026-03-17 16:05:00.746138733 +0000 UTC m=+1692.159454780" lastFinishedPulling="2026-03-17 16:05:42.909226132 +0000 UTC m=+1734.322542179" observedRunningTime="2026-03-17 16:05:43.213483205 +0000 UTC m=+1734.626799252" watchObservedRunningTime="2026-03-17 16:05:43.217325085 +0000 UTC m=+1734.630641132" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.322290 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-658d967468-8hprg"] Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.339222 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-658d967468-8hprg"] Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.371000 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00a982e7-017e-4f3c-a8aa-21eab359cad4" path="/var/lib/kubelet/pods/00a982e7-017e-4f3c-a8aa-21eab359cad4/volumes" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.374193 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tfkg\" (UniqueName: \"kubernetes.io/projected/3caf16d7-044a-4efc-a3ba-c494255df8ff-kube-api-access-7tfkg\") pod \"3caf16d7-044a-4efc-a3ba-c494255df8ff\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.374243 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data-custom\") pod \"3caf16d7-044a-4efc-a3ba-c494255df8ff\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.374380 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-combined-ca-bundle\") pod \"3caf16d7-044a-4efc-a3ba-c494255df8ff\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.374455 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data\") pod \"3caf16d7-044a-4efc-a3ba-c494255df8ff\" (UID: \"3caf16d7-044a-4efc-a3ba-c494255df8ff\") " Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.380865 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3caf16d7-044a-4efc-a3ba-c494255df8ff" (UID: "3caf16d7-044a-4efc-a3ba-c494255df8ff"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.381566 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3caf16d7-044a-4efc-a3ba-c494255df8ff-kube-api-access-7tfkg" (OuterVolumeSpecName: "kube-api-access-7tfkg") pod "3caf16d7-044a-4efc-a3ba-c494255df8ff" (UID: "3caf16d7-044a-4efc-a3ba-c494255df8ff"). InnerVolumeSpecName "kube-api-access-7tfkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.449271 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3caf16d7-044a-4efc-a3ba-c494255df8ff" (UID: "3caf16d7-044a-4efc-a3ba-c494255df8ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.470274 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data" (OuterVolumeSpecName: "config-data") pod "3caf16d7-044a-4efc-a3ba-c494255df8ff" (UID: "3caf16d7-044a-4efc-a3ba-c494255df8ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.479164 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.479258 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tfkg\" (UniqueName: \"kubernetes.io/projected/3caf16d7-044a-4efc-a3ba-c494255df8ff-kube-api-access-7tfkg\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.479272 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:43 crc kubenswrapper[4767]: I0317 16:05:43.479283 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3caf16d7-044a-4efc-a3ba-c494255df8ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:44 crc kubenswrapper[4767]: I0317 16:05:44.210823 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-75777db76d-5z4qz" event={"ID":"3caf16d7-044a-4efc-a3ba-c494255df8ff","Type":"ContainerDied","Data":"12e3c38bada1455b32016242209dff4f42ea450915eb2bd79df88618a4bd864a"} Mar 17 16:05:44 crc kubenswrapper[4767]: I0317 16:05:44.211248 4767 scope.go:117] "RemoveContainer" containerID="b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e" Mar 17 16:05:44 crc kubenswrapper[4767]: I0317 16:05:44.210893 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-75777db76d-5z4qz" Mar 17 16:05:44 crc kubenswrapper[4767]: I0317 16:05:44.267970 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-75777db76d-5z4qz"] Mar 17 16:05:44 crc kubenswrapper[4767]: I0317 16:05:44.285194 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-75777db76d-5z4qz"] Mar 17 16:05:44 crc kubenswrapper[4767]: I0317 16:05:44.473990 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 17 16:05:44 crc kubenswrapper[4767]: I0317 16:05:44.474363 4767 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 16:05:44 crc kubenswrapper[4767]: I0317 16:05:44.477255 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 17 16:05:45 crc kubenswrapper[4767]: I0317 16:05:45.382464 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3caf16d7-044a-4efc-a3ba-c494255df8ff" path="/var/lib/kubelet/pods/3caf16d7-044a-4efc-a3ba-c494255df8ff/volumes" Mar 17 16:05:46 crc kubenswrapper[4767]: I0317 16:05:46.355595 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:05:46 crc kubenswrapper[4767]: E0317 16:05:46.356476 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:05:55 crc kubenswrapper[4767]: I0317 16:05:55.680021 4767 generic.go:334] "Generic (PLEG): container finished" podID="dba94162-c0cc-46e7-a4ca-45836a003a08" containerID="19902911191b266a4117332e9664ffbe0b11a0de4db4a9275ffc9802b5c13c9e" exitCode=0 Mar 17 16:05:55 crc kubenswrapper[4767]: I0317 16:05:55.680090 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dsn5f" event={"ID":"dba94162-c0cc-46e7-a4ca-45836a003a08","Type":"ContainerDied","Data":"19902911191b266a4117332e9664ffbe0b11a0de4db4a9275ffc9802b5c13c9e"} Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.601630 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.761548 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-combined-ca-bundle\") pod \"dba94162-c0cc-46e7-a4ca-45836a003a08\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.761637 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggltx\" (UniqueName: \"kubernetes.io/projected/dba94162-c0cc-46e7-a4ca-45836a003a08-kube-api-access-ggltx\") pod \"dba94162-c0cc-46e7-a4ca-45836a003a08\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.761819 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-config-data\") pod \"dba94162-c0cc-46e7-a4ca-45836a003a08\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.762037 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-scripts\") pod \"dba94162-c0cc-46e7-a4ca-45836a003a08\" (UID: \"dba94162-c0cc-46e7-a4ca-45836a003a08\") " Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.771991 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-scripts" (OuterVolumeSpecName: "scripts") pod "dba94162-c0cc-46e7-a4ca-45836a003a08" (UID: "dba94162-c0cc-46e7-a4ca-45836a003a08"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.782722 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dba94162-c0cc-46e7-a4ca-45836a003a08-kube-api-access-ggltx" (OuterVolumeSpecName: "kube-api-access-ggltx") pod "dba94162-c0cc-46e7-a4ca-45836a003a08" (UID: "dba94162-c0cc-46e7-a4ca-45836a003a08"). InnerVolumeSpecName "kube-api-access-ggltx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.795732 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dba94162-c0cc-46e7-a4ca-45836a003a08" (UID: "dba94162-c0cc-46e7-a4ca-45836a003a08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.805388 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-config-data" (OuterVolumeSpecName: "config-data") pod "dba94162-c0cc-46e7-a4ca-45836a003a08" (UID: "dba94162-c0cc-46e7-a4ca-45836a003a08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.864770 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.864809 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.864817 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba94162-c0cc-46e7-a4ca-45836a003a08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:57 crc kubenswrapper[4767]: I0317 16:05:57.864830 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggltx\" (UniqueName: \"kubernetes.io/projected/dba94162-c0cc-46e7-a4ca-45836a003a08-kube-api-access-ggltx\") on node \"crc\" DevicePath \"\"" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.012358 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dsn5f" event={"ID":"dba94162-c0cc-46e7-a4ca-45836a003a08","Type":"ContainerDied","Data":"0cfe300fc7f0e3a19680132d190529f6b8a6c4da47e533665e133efe3d1cd005"} Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.012412 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cfe300fc7f0e3a19680132d190529f6b8a6c4da47e533665e133efe3d1cd005" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.012436 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dsn5f" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.068219 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 17 16:05:58 crc kubenswrapper[4767]: E0317 16:05:58.068830 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65a045e9-219c-4432-9d80-09a944c85013" containerName="extract-utilities" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.068856 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="65a045e9-219c-4432-9d80-09a944c85013" containerName="extract-utilities" Mar 17 16:05:58 crc kubenswrapper[4767]: E0317 16:05:58.068892 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dba94162-c0cc-46e7-a4ca-45836a003a08" containerName="nova-cell0-conductor-db-sync" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.068899 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="dba94162-c0cc-46e7-a4ca-45836a003a08" containerName="nova-cell0-conductor-db-sync" Mar 17 16:05:58 crc kubenswrapper[4767]: E0317 16:05:58.068914 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65a045e9-219c-4432-9d80-09a944c85013" containerName="registry-server" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.068921 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="65a045e9-219c-4432-9d80-09a944c85013" containerName="registry-server" Mar 17 16:05:58 crc kubenswrapper[4767]: E0317 16:05:58.068930 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65a045e9-219c-4432-9d80-09a944c85013" containerName="extract-content" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.068936 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="65a045e9-219c-4432-9d80-09a944c85013" containerName="extract-content" Mar 17 16:05:58 crc kubenswrapper[4767]: E0317 16:05:58.068948 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a982e7-017e-4f3c-a8aa-21eab359cad4" containerName="heat-cfnapi" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.068954 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a982e7-017e-4f3c-a8aa-21eab359cad4" containerName="heat-cfnapi" Mar 17 16:05:58 crc kubenswrapper[4767]: E0317 16:05:58.068966 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3caf16d7-044a-4efc-a3ba-c494255df8ff" containerName="heat-api" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.068973 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3caf16d7-044a-4efc-a3ba-c494255df8ff" containerName="heat-api" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.069211 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="dba94162-c0cc-46e7-a4ca-45836a003a08" containerName="nova-cell0-conductor-db-sync" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.069229 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="65a045e9-219c-4432-9d80-09a944c85013" containerName="registry-server" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.069251 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a982e7-017e-4f3c-a8aa-21eab359cad4" containerName="heat-cfnapi" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.069260 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3caf16d7-044a-4efc-a3ba-c494255df8ff" containerName="heat-api" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.070580 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.082934 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.083085 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-kp9tf" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.088670 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.172314 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/987da0c6-fab3-4111-a6fc-174fed6b6f80-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"987da0c6-fab3-4111-a6fc-174fed6b6f80\") " pod="openstack/nova-cell0-conductor-0" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.172703 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/987da0c6-fab3-4111-a6fc-174fed6b6f80-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"987da0c6-fab3-4111-a6fc-174fed6b6f80\") " pod="openstack/nova-cell0-conductor-0" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.173003 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88v5m\" (UniqueName: \"kubernetes.io/projected/987da0c6-fab3-4111-a6fc-174fed6b6f80-kube-api-access-88v5m\") pod \"nova-cell0-conductor-0\" (UID: \"987da0c6-fab3-4111-a6fc-174fed6b6f80\") " pod="openstack/nova-cell0-conductor-0" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.275659 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/987da0c6-fab3-4111-a6fc-174fed6b6f80-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"987da0c6-fab3-4111-a6fc-174fed6b6f80\") " pod="openstack/nova-cell0-conductor-0" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.275965 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88v5m\" (UniqueName: \"kubernetes.io/projected/987da0c6-fab3-4111-a6fc-174fed6b6f80-kube-api-access-88v5m\") pod \"nova-cell0-conductor-0\" (UID: \"987da0c6-fab3-4111-a6fc-174fed6b6f80\") " pod="openstack/nova-cell0-conductor-0" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.276129 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/987da0c6-fab3-4111-a6fc-174fed6b6f80-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"987da0c6-fab3-4111-a6fc-174fed6b6f80\") " pod="openstack/nova-cell0-conductor-0" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.287079 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/987da0c6-fab3-4111-a6fc-174fed6b6f80-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"987da0c6-fab3-4111-a6fc-174fed6b6f80\") " pod="openstack/nova-cell0-conductor-0" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.300157 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/987da0c6-fab3-4111-a6fc-174fed6b6f80-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"987da0c6-fab3-4111-a6fc-174fed6b6f80\") " pod="openstack/nova-cell0-conductor-0" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.308475 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88v5m\" (UniqueName: \"kubernetes.io/projected/987da0c6-fab3-4111-a6fc-174fed6b6f80-kube-api-access-88v5m\") pod \"nova-cell0-conductor-0\" (UID: \"987da0c6-fab3-4111-a6fc-174fed6b6f80\") " pod="openstack/nova-cell0-conductor-0" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.402221 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 17 16:05:58 crc kubenswrapper[4767]: I0317 16:05:58.995526 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 17 16:05:59 crc kubenswrapper[4767]: I0317 16:05:59.026533 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"987da0c6-fab3-4111-a6fc-174fed6b6f80","Type":"ContainerStarted","Data":"df2e8acad1966b7bfa3f0769f515f561c3ea6fc5ffae43e570926226a3050ec4"} Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.021950 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.040643 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"987da0c6-fab3-4111-a6fc-174fed6b6f80","Type":"ContainerStarted","Data":"f48b17bc5857069bd416f625d54daa87a6ebb8a4b35de8cb92a5ff744c4c0d09"} Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.041108 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.065115 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.065092747 podStartE2EDuration="2.065092747s" podCreationTimestamp="2026-03-17 16:05:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:06:00.059859377 +0000 UTC m=+1751.473175434" watchObservedRunningTime="2026-03-17 16:06:00.065092747 +0000 UTC m=+1751.478408794" Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.150924 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562726-sq886"] Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.153341 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562726-sq886" Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.156137 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.157962 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.159699 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.167740 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562726-sq886"] Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.607345 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:06:00 crc kubenswrapper[4767]: E0317 16:06:00.607932 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.613355 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz6z9\" (UniqueName: \"kubernetes.io/projected/54cf85c2-ceca-418b-a22e-ecadc2a3acfc-kube-api-access-bz6z9\") pod \"auto-csr-approver-29562726-sq886\" (UID: \"54cf85c2-ceca-418b-a22e-ecadc2a3acfc\") " pod="openshift-infra/auto-csr-approver-29562726-sq886" Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.721260 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz6z9\" (UniqueName: \"kubernetes.io/projected/54cf85c2-ceca-418b-a22e-ecadc2a3acfc-kube-api-access-bz6z9\") pod \"auto-csr-approver-29562726-sq886\" (UID: \"54cf85c2-ceca-418b-a22e-ecadc2a3acfc\") " pod="openshift-infra/auto-csr-approver-29562726-sq886" Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.746684 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz6z9\" (UniqueName: \"kubernetes.io/projected/54cf85c2-ceca-418b-a22e-ecadc2a3acfc-kube-api-access-bz6z9\") pod \"auto-csr-approver-29562726-sq886\" (UID: \"54cf85c2-ceca-418b-a22e-ecadc2a3acfc\") " pod="openshift-infra/auto-csr-approver-29562726-sq886" Mar 17 16:06:00 crc kubenswrapper[4767]: I0317 16:06:00.792572 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562726-sq886" Mar 17 16:06:01 crc kubenswrapper[4767]: I0317 16:06:01.421494 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562726-sq886"] Mar 17 16:06:02 crc kubenswrapper[4767]: I0317 16:06:02.087544 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562726-sq886" event={"ID":"54cf85c2-ceca-418b-a22e-ecadc2a3acfc","Type":"ContainerStarted","Data":"5f8a1904b9c7a4ed11cc5719151da41a160eee25290097e20adbd1d97bc44f4b"} Mar 17 16:06:04 crc kubenswrapper[4767]: I0317 16:06:04.291473 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562726-sq886" event={"ID":"54cf85c2-ceca-418b-a22e-ecadc2a3acfc","Type":"ContainerStarted","Data":"e9a43787ac418a99df162e2c9d652dac1b1aecc999986ccde30a0737758453e1"} Mar 17 16:06:04 crc kubenswrapper[4767]: I0317 16:06:04.321761 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562726-sq886" podStartSLOduration=2.855857178 podStartE2EDuration="4.321735394s" podCreationTimestamp="2026-03-17 16:06:00 +0000 UTC" firstStartedPulling="2026-03-17 16:06:01.401180892 +0000 UTC m=+1752.814496939" lastFinishedPulling="2026-03-17 16:06:02.867059108 +0000 UTC m=+1754.280375155" observedRunningTime="2026-03-17 16:06:04.316539325 +0000 UTC m=+1755.729855382" watchObservedRunningTime="2026-03-17 16:06:04.321735394 +0000 UTC m=+1755.735051441" Mar 17 16:06:05 crc kubenswrapper[4767]: I0317 16:06:05.307277 4767 generic.go:334] "Generic (PLEG): container finished" podID="54cf85c2-ceca-418b-a22e-ecadc2a3acfc" containerID="e9a43787ac418a99df162e2c9d652dac1b1aecc999986ccde30a0737758453e1" exitCode=0 Mar 17 16:06:05 crc kubenswrapper[4767]: I0317 16:06:05.307363 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562726-sq886" event={"ID":"54cf85c2-ceca-418b-a22e-ecadc2a3acfc","Type":"ContainerDied","Data":"e9a43787ac418a99df162e2c9d652dac1b1aecc999986ccde30a0737758453e1"} Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.071230 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562726-sq886" Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.171982 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bz6z9\" (UniqueName: \"kubernetes.io/projected/54cf85c2-ceca-418b-a22e-ecadc2a3acfc-kube-api-access-bz6z9\") pod \"54cf85c2-ceca-418b-a22e-ecadc2a3acfc\" (UID: \"54cf85c2-ceca-418b-a22e-ecadc2a3acfc\") " Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.180515 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54cf85c2-ceca-418b-a22e-ecadc2a3acfc-kube-api-access-bz6z9" (OuterVolumeSpecName: "kube-api-access-bz6z9") pod "54cf85c2-ceca-418b-a22e-ecadc2a3acfc" (UID: "54cf85c2-ceca-418b-a22e-ecadc2a3acfc"). InnerVolumeSpecName "kube-api-access-bz6z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.275534 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bz6z9\" (UniqueName: \"kubernetes.io/projected/54cf85c2-ceca-418b-a22e-ecadc2a3acfc-kube-api-access-bz6z9\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:07 crc kubenswrapper[4767]: E0317 16:06:07.352296 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65f3e07f_47cf_4973_82e9_353692ec4a0b.slice/crio-7088ba7b37e5f96402f31a68ef2fbaea4383aa9aee7ec624b38e6cec0dd49c64\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddba94162_c0cc_46e7_a4ca_45836a003a08.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-12e3c38bada1455b32016242209dff4f42ea450915eb2bd79df88618a4bd864a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-conmon-b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddba94162_c0cc_46e7_a4ca_45836a003a08.slice/crio-0cfe300fc7f0e3a19680132d190529f6b8a6c4da47e533665e133efe3d1cd005\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-f646134bf7c3e81d39895e638d392ddf0d379219ae5e27f819f2a8aecd6a8030\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:07 crc kubenswrapper[4767]: E0317 16:06:07.354753 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-conmon-b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddba94162_c0cc_46e7_a4ca_45836a003a08.slice/crio-0cfe300fc7f0e3a19680132d190529f6b8a6c4da47e533665e133efe3d1cd005\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddba94162_c0cc_46e7_a4ca_45836a003a08.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-12e3c38bada1455b32016242209dff4f42ea450915eb2bd79df88618a4bd864a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-f646134bf7c3e81d39895e638d392ddf0d379219ae5e27f819f2a8aecd6a8030\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:07 crc kubenswrapper[4767]: E0317 16:06:07.356508 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-f646134bf7c3e81d39895e638d392ddf0d379219ae5e27f819f2a8aecd6a8030\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-conmon-b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddba94162_c0cc_46e7_a4ca_45836a003a08.slice/crio-0cfe300fc7f0e3a19680132d190529f6b8a6c4da47e533665e133efe3d1cd005\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddba94162_c0cc_46e7_a4ca_45836a003a08.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-12e3c38bada1455b32016242209dff4f42ea450915eb2bd79df88618a4bd864a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:07 crc kubenswrapper[4767]: E0317 16:06:07.374692 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice/crio-f646134bf7c3e81d39895e638d392ddf0d379219ae5e27f819f2a8aecd6a8030\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-12e3c38bada1455b32016242209dff4f42ea450915eb2bd79df88618a4bd864a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a982e7_017e_4f3c_a8aa_21eab359cad4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice/crio-conmon-b46b7595173bfaf96f0b38b48bfcf16973e565580d7a107589c71755b6e7949e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3caf16d7_044a_4efc_a3ba_c494255df8ff.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddba94162_c0cc_46e7_a4ca_45836a003a08.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.383258 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562726-sq886" Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.395240 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562726-sq886" event={"ID":"54cf85c2-ceca-418b-a22e-ecadc2a3acfc","Type":"ContainerDied","Data":"5f8a1904b9c7a4ed11cc5719151da41a160eee25290097e20adbd1d97bc44f4b"} Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.395335 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f8a1904b9c7a4ed11cc5719151da41a160eee25290097e20adbd1d97bc44f4b" Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.400471 4767 generic.go:334] "Generic (PLEG): container finished" podID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerID="bffa8f087ec5113fb13fa1cd880317f4f67679bd0724b257a717a09ba28c4cb2" exitCode=137 Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.400541 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65f3e07f-47cf-4973-82e9-353692ec4a0b","Type":"ContainerDied","Data":"bffa8f087ec5113fb13fa1cd880317f4f67679bd0724b257a717a09ba28c4cb2"} Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.513939 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562720-bgk8g"] Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.562683 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562720-bgk8g"] Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.751318 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.905979 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75dg6\" (UniqueName: \"kubernetes.io/projected/65f3e07f-47cf-4973-82e9-353692ec4a0b-kube-api-access-75dg6\") pod \"65f3e07f-47cf-4973-82e9-353692ec4a0b\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.906110 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-combined-ca-bundle\") pod \"65f3e07f-47cf-4973-82e9-353692ec4a0b\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.906346 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-log-httpd\") pod \"65f3e07f-47cf-4973-82e9-353692ec4a0b\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.906402 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-sg-core-conf-yaml\") pod \"65f3e07f-47cf-4973-82e9-353692ec4a0b\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.906538 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-run-httpd\") pod \"65f3e07f-47cf-4973-82e9-353692ec4a0b\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.906575 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-config-data\") pod \"65f3e07f-47cf-4973-82e9-353692ec4a0b\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.906618 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-scripts\") pod \"65f3e07f-47cf-4973-82e9-353692ec4a0b\" (UID: \"65f3e07f-47cf-4973-82e9-353692ec4a0b\") " Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.907825 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "65f3e07f-47cf-4973-82e9-353692ec4a0b" (UID: "65f3e07f-47cf-4973-82e9-353692ec4a0b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.908387 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "65f3e07f-47cf-4973-82e9-353692ec4a0b" (UID: "65f3e07f-47cf-4973-82e9-353692ec4a0b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.915541 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-scripts" (OuterVolumeSpecName: "scripts") pod "65f3e07f-47cf-4973-82e9-353692ec4a0b" (UID: "65f3e07f-47cf-4973-82e9-353692ec4a0b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.915594 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65f3e07f-47cf-4973-82e9-353692ec4a0b-kube-api-access-75dg6" (OuterVolumeSpecName: "kube-api-access-75dg6") pod "65f3e07f-47cf-4973-82e9-353692ec4a0b" (UID: "65f3e07f-47cf-4973-82e9-353692ec4a0b"). InnerVolumeSpecName "kube-api-access-75dg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:07 crc kubenswrapper[4767]: I0317 16:06:07.945491 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "65f3e07f-47cf-4973-82e9-353692ec4a0b" (UID: "65f3e07f-47cf-4973-82e9-353692ec4a0b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.009804 4767 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.009847 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.009861 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75dg6\" (UniqueName: \"kubernetes.io/projected/65f3e07f-47cf-4973-82e9-353692ec4a0b-kube-api-access-75dg6\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.009873 4767 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65f3e07f-47cf-4973-82e9-353692ec4a0b-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.009885 4767 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.058453 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65f3e07f-47cf-4973-82e9-353692ec4a0b" (UID: "65f3e07f-47cf-4973-82e9-353692ec4a0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.080210 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-config-data" (OuterVolumeSpecName: "config-data") pod "65f3e07f-47cf-4973-82e9-353692ec4a0b" (UID: "65f3e07f-47cf-4973-82e9-353692ec4a0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.112808 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.112874 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f3e07f-47cf-4973-82e9-353692ec4a0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.468215 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65f3e07f-47cf-4973-82e9-353692ec4a0b","Type":"ContainerDied","Data":"7088ba7b37e5f96402f31a68ef2fbaea4383aa9aee7ec624b38e6cec0dd49c64"} Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.468286 4767 scope.go:117] "RemoveContainer" containerID="bffa8f087ec5113fb13fa1cd880317f4f67679bd0724b257a717a09ba28c4cb2" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.468432 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.476692 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.530047 4767 scope.go:117] "RemoveContainer" containerID="61fa65eb2e6a84d997f3520cb68e39294a7eb109dfff6c4bdd37368a839d2767" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.542361 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.551538 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.595941 4767 scope.go:117] "RemoveContainer" containerID="ed6d5c7f2bdbef020415709a1c9d8bba43d9bf3780d06686c9f18e737513854c" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.606726 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:08 crc kubenswrapper[4767]: E0317 16:06:08.607654 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="ceilometer-notification-agent" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.607692 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="ceilometer-notification-agent" Mar 17 16:06:08 crc kubenswrapper[4767]: E0317 16:06:08.607708 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cf85c2-ceca-418b-a22e-ecadc2a3acfc" containerName="oc" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.607717 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cf85c2-ceca-418b-a22e-ecadc2a3acfc" containerName="oc" Mar 17 16:06:08 crc kubenswrapper[4767]: E0317 16:06:08.607752 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="ceilometer-central-agent" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.607759 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="ceilometer-central-agent" Mar 17 16:06:08 crc kubenswrapper[4767]: E0317 16:06:08.607776 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="proxy-httpd" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.607785 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="proxy-httpd" Mar 17 16:06:08 crc kubenswrapper[4767]: E0317 16:06:08.607803 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="sg-core" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.607811 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="sg-core" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.608119 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="proxy-httpd" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.608149 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cf85c2-ceca-418b-a22e-ecadc2a3acfc" containerName="oc" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.608162 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="sg-core" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.608978 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="ceilometer-central-agent" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.609004 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" containerName="ceilometer-notification-agent" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.611546 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.618066 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.618412 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.653698 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.680694 4767 scope.go:117] "RemoveContainer" containerID="02333d1ad02cc38464ce2d996650bc5058ce4ce600c1ffaeb12bc4deee18a0c3" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.741194 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-run-httpd\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.741286 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.741336 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-log-httpd\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.741410 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqmq9\" (UniqueName: \"kubernetes.io/projected/3586e791-e20d-44ea-a784-fd7c14389087-kube-api-access-xqmq9\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.741449 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-scripts\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.741527 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.741657 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-config-data\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.844296 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-run-httpd\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.844379 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.844420 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-log-httpd\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.844482 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqmq9\" (UniqueName: \"kubernetes.io/projected/3586e791-e20d-44ea-a784-fd7c14389087-kube-api-access-xqmq9\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.844511 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-scripts\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.844550 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.844635 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-config-data\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.847143 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-log-httpd\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.847495 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-run-httpd\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.851424 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.851705 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-config-data\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.851601 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-scripts\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.875470 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.880575 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqmq9\" (UniqueName: \"kubernetes.io/projected/3586e791-e20d-44ea-a784-fd7c14389087-kube-api-access-xqmq9\") pod \"ceilometer-0\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.944588 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.980852 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-nlvws"] Mar 17 16:06:08 crc kubenswrapper[4767]: I0317 16:06:08.983103 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-nlvws" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.003736 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-nlvws"] Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.060696 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7f88\" (UniqueName: \"kubernetes.io/projected/46218023-25a1-49e3-8358-08445caf00ce-kube-api-access-t7f88\") pod \"aodh-db-create-nlvws\" (UID: \"46218023-25a1-49e3-8358-08445caf00ce\") " pod="openstack/aodh-db-create-nlvws" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.060795 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46218023-25a1-49e3-8358-08445caf00ce-operator-scripts\") pod \"aodh-db-create-nlvws\" (UID: \"46218023-25a1-49e3-8358-08445caf00ce\") " pod="openstack/aodh-db-create-nlvws" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.092256 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-21c1-account-create-update-2g64g"] Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.095002 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-21c1-account-create-update-2g64g" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.100533 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.108369 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-21c1-account-create-update-2g64g"] Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.164561 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7f88\" (UniqueName: \"kubernetes.io/projected/46218023-25a1-49e3-8358-08445caf00ce-kube-api-access-t7f88\") pod \"aodh-db-create-nlvws\" (UID: \"46218023-25a1-49e3-8358-08445caf00ce\") " pod="openstack/aodh-db-create-nlvws" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.164977 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46218023-25a1-49e3-8358-08445caf00ce-operator-scripts\") pod \"aodh-db-create-nlvws\" (UID: \"46218023-25a1-49e3-8358-08445caf00ce\") " pod="openstack/aodh-db-create-nlvws" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.166446 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46218023-25a1-49e3-8358-08445caf00ce-operator-scripts\") pod \"aodh-db-create-nlvws\" (UID: \"46218023-25a1-49e3-8358-08445caf00ce\") " pod="openstack/aodh-db-create-nlvws" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.190873 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7f88\" (UniqueName: \"kubernetes.io/projected/46218023-25a1-49e3-8358-08445caf00ce-kube-api-access-t7f88\") pod \"aodh-db-create-nlvws\" (UID: \"46218023-25a1-49e3-8358-08445caf00ce\") " pod="openstack/aodh-db-create-nlvws" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.268035 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/271cc3d9-d20c-4708-bd6a-765a8c439cf6-operator-scripts\") pod \"aodh-21c1-account-create-update-2g64g\" (UID: \"271cc3d9-d20c-4708-bd6a-765a8c439cf6\") " pod="openstack/aodh-21c1-account-create-update-2g64g" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.268580 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfbrh\" (UniqueName: \"kubernetes.io/projected/271cc3d9-d20c-4708-bd6a-765a8c439cf6-kube-api-access-mfbrh\") pod \"aodh-21c1-account-create-update-2g64g\" (UID: \"271cc3d9-d20c-4708-bd6a-765a8c439cf6\") " pod="openstack/aodh-21c1-account-create-update-2g64g" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.373830 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/271cc3d9-d20c-4708-bd6a-765a8c439cf6-operator-scripts\") pod \"aodh-21c1-account-create-update-2g64g\" (UID: \"271cc3d9-d20c-4708-bd6a-765a8c439cf6\") " pod="openstack/aodh-21c1-account-create-update-2g64g" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.374454 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfbrh\" (UniqueName: \"kubernetes.io/projected/271cc3d9-d20c-4708-bd6a-765a8c439cf6-kube-api-access-mfbrh\") pod \"aodh-21c1-account-create-update-2g64g\" (UID: \"271cc3d9-d20c-4708-bd6a-765a8c439cf6\") " pod="openstack/aodh-21c1-account-create-update-2g64g" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.374877 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/271cc3d9-d20c-4708-bd6a-765a8c439cf6-operator-scripts\") pod \"aodh-21c1-account-create-update-2g64g\" (UID: \"271cc3d9-d20c-4708-bd6a-765a8c439cf6\") " pod="openstack/aodh-21c1-account-create-update-2g64g" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.403058 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65f3e07f-47cf-4973-82e9-353692ec4a0b" path="/var/lib/kubelet/pods/65f3e07f-47cf-4973-82e9-353692ec4a0b/volumes" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.405131 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93755875-2f62-4a18-aa46-a2c55496e1c2" path="/var/lib/kubelet/pods/93755875-2f62-4a18-aa46-a2c55496e1c2/volumes" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.418115 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-nlvws" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.452194 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfbrh\" (UniqueName: \"kubernetes.io/projected/271cc3d9-d20c-4708-bd6a-765a8c439cf6-kube-api-access-mfbrh\") pod \"aodh-21c1-account-create-update-2g64g\" (UID: \"271cc3d9-d20c-4708-bd6a-765a8c439cf6\") " pod="openstack/aodh-21c1-account-create-update-2g64g" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.547526 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-p7cb5"] Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.552183 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.558480 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.558829 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.682292 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-p7cb5"] Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.700264 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.700350 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-config-data\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.700486 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-scripts\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.700762 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cznlf\" (UniqueName: \"kubernetes.io/projected/0aaae974-d49d-4980-a540-6bc2ec85a15f-kube-api-access-cznlf\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.714780 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.738883 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-21c1-account-create-update-2g64g" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.824135 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.833761 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-config-data\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.834051 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-scripts\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.834767 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cznlf\" (UniqueName: \"kubernetes.io/projected/0aaae974-d49d-4980-a540-6bc2ec85a15f-kube-api-access-cznlf\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.873538 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-config-data\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.874138 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.875541 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-scripts\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.913970 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cznlf\" (UniqueName: \"kubernetes.io/projected/0aaae974-d49d-4980-a540-6bc2ec85a15f-kube-api-access-cznlf\") pod \"nova-cell0-cell-mapping-p7cb5\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:09 crc kubenswrapper[4767]: I0317 16:06:09.975704 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.074217 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.076756 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.082307 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.174913 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw888\" (UniqueName: \"kubernetes.io/projected/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-kube-api-access-xw888\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.175334 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-logs\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.175425 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.175604 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-config-data\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.191968 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.195595 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.203319 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.206604 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.280562 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw888\" (UniqueName: \"kubernetes.io/projected/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-kube-api-access-xw888\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.280723 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-logs\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.280777 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.280928 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-config-data\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.282372 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-logs\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.287138 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.308730 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.344112 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-config-data\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.344112 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.367784 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.369989 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.378004 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.385548 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-config-data\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.385703 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjwk2\" (UniqueName: \"kubernetes.io/projected/0fb28438-621e-41a8-b1ad-a07bf5e181c5-kube-api-access-cjwk2\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.385749 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.386133 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb28438-621e-41a8-b1ad-a07bf5e181c5-logs\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.402438 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw888\" (UniqueName: \"kubernetes.io/projected/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-kube-api-access-xw888\") pod \"nova-api-0\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.405365 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.435897 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.457953 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-z9szv"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.464367 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.503102 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjwk2\" (UniqueName: \"kubernetes.io/projected/0fb28438-621e-41a8-b1ad-a07bf5e181c5-kube-api-access-cjwk2\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.503190 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.503346 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.503414 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.503511 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-config-data\") pod \"nova-scheduler-0\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.503565 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb28438-621e-41a8-b1ad-a07bf5e181c5-logs\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.503640 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkjwf\" (UniqueName: \"kubernetes.io/projected/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-kube-api-access-wkjwf\") pod \"nova-scheduler-0\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.503759 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-config-data\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.520500 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb28438-621e-41a8-b1ad-a07bf5e181c5-logs\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.523395 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.527792 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.531890 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.540913 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-config-data\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.579868 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjwk2\" (UniqueName: \"kubernetes.io/projected/0fb28438-621e-41a8-b1ad-a07bf5e181c5-kube-api-access-cjwk2\") pod \"nova-metadata-0\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.591263 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.608933 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrvp8\" (UniqueName: \"kubernetes.io/projected/f3467dd4-3ba8-4492-a678-611b6b982946-kube-api-access-zrvp8\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.609047 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-config-data\") pod \"nova-scheduler-0\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.609099 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.609162 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.609248 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-svc\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.609316 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.609363 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkjwf\" (UniqueName: \"kubernetes.io/projected/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-kube-api-access-wkjwf\") pod \"nova-scheduler-0\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.609389 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.609506 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8sz2\" (UniqueName: \"kubernetes.io/projected/64c09859-d250-40a0-97a1-aa8fd1371432-kube-api-access-z8sz2\") pod \"nova-cell1-novncproxy-0\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.609635 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-config\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.609687 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.609908 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.635830 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.644854 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-config-data\") pod \"nova-scheduler-0\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.667314 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-z9szv"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.669405 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkjwf\" (UniqueName: \"kubernetes.io/projected/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-kube-api-access-wkjwf\") pod \"nova-scheduler-0\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.680028 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3586e791-e20d-44ea-a784-fd7c14389087","Type":"ContainerStarted","Data":"a7c2f8e0ca758482fa73a45c18b5ae75123121ff00219b9ffa0aad58960f77fa"} Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.713988 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrvp8\" (UniqueName: \"kubernetes.io/projected/f3467dd4-3ba8-4492-a678-611b6b982946-kube-api-access-zrvp8\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.714083 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.714120 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.714149 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-svc\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.714198 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.714224 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.714277 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8sz2\" (UniqueName: \"kubernetes.io/projected/64c09859-d250-40a0-97a1-aa8fd1371432-kube-api-access-z8sz2\") pod \"nova-cell1-novncproxy-0\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.714340 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-config\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.714367 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.715548 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.719230 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.727157 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.727862 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-svc\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.728415 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.729471 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-config\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.733384 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.757113 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrvp8\" (UniqueName: \"kubernetes.io/projected/f3467dd4-3ba8-4492-a678-611b6b982946-kube-api-access-zrvp8\") pod \"dnsmasq-dns-9b86998b5-z9szv\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.792221 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8sz2\" (UniqueName: \"kubernetes.io/projected/64c09859-d250-40a0-97a1-aa8fd1371432-kube-api-access-z8sz2\") pod \"nova-cell1-novncproxy-0\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.823086 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.853107 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.854210 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.941317 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-nlvws"] Mar 17 16:06:10 crc kubenswrapper[4767]: I0317 16:06:10.943489 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:11 crc kubenswrapper[4767]: I0317 16:06:11.120533 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-21c1-account-create-update-2g64g"] Mar 17 16:06:11 crc kubenswrapper[4767]: I0317 16:06:11.906845 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-nlvws" event={"ID":"46218023-25a1-49e3-8358-08445caf00ce","Type":"ContainerStarted","Data":"53e90e2cc857a3744cee05dad708820124c926eb163cd2b1320be31f59d979a8"} Mar 17 16:06:11 crc kubenswrapper[4767]: I0317 16:06:11.928838 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-21c1-account-create-update-2g64g" event={"ID":"271cc3d9-d20c-4708-bd6a-765a8c439cf6","Type":"ContainerStarted","Data":"dad31ae592e15fa8c1a179afcc2356f6d6a46f73f4bf1c15bd5d1afcf698a27a"} Mar 17 16:06:11 crc kubenswrapper[4767]: I0317 16:06:11.956869 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-p7cb5"] Mar 17 16:06:12 crc kubenswrapper[4767]: I0317 16:06:12.484039 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:06:12 crc kubenswrapper[4767]: I0317 16:06:12.804611 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-z9szv"] Mar 17 16:06:12 crc kubenswrapper[4767]: I0317 16:06:12.901772 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:06:12 crc kubenswrapper[4767]: I0317 16:06:12.979843 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-21c1-account-create-update-2g64g" event={"ID":"271cc3d9-d20c-4708-bd6a-765a8c439cf6","Type":"ContainerStarted","Data":"eea4f2a1ea128dac6e5fa290ffcaf4f7e32b29d8e34f63c19f28c0eebd49d236"} Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.009560 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c60ded0-ef6b-4b07-93ef-b0aa6c643084","Type":"ContainerStarted","Data":"6f3a64a53fad6d3b105ce6775ffee382fe9512563d49e10b7e85eee2429ce596"} Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.035096 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9fv46"] Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.036996 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.038657 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-p7cb5" event={"ID":"0aaae974-d49d-4980-a540-6bc2ec85a15f","Type":"ContainerStarted","Data":"466e5152b4dae52e57a86166113f0ef3d5745577ca45e1f223ca9a7f0065c3f8"} Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.038684 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-p7cb5" event={"ID":"0aaae974-d49d-4980-a540-6bc2ec85a15f","Type":"ContainerStarted","Data":"973b7ace9ff4adc22087bd2be6b8d9340ebab4ddd6ce94fe8108f9f8bb7014ad"} Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.047476 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.048444 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.084756 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3586e791-e20d-44ea-a784-fd7c14389087","Type":"ContainerStarted","Data":"ec289db6fac44df87d03cf5f2520718277113a8a3b75ae87ec8a5c7ec99c7184"} Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.102751 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9fv46"] Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.125489 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" event={"ID":"f3467dd4-3ba8-4492-a678-611b6b982946","Type":"ContainerStarted","Data":"47263d72b689335533cdad3559966dedcda33ce22076eb1ddf963a2c426433b7"} Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.128911 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc","Type":"ContainerStarted","Data":"d3b6e2e96e0298b68e7e30bcee021aaa958092f32cd0e18fd0841fd07ba20124"} Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.144687 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-21c1-account-create-update-2g64g" podStartSLOduration=4.144655664 podStartE2EDuration="4.144655664s" podCreationTimestamp="2026-03-17 16:06:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:06:13.023310513 +0000 UTC m=+1764.436626560" watchObservedRunningTime="2026-03-17 16:06:13.144655664 +0000 UTC m=+1764.557971731" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.150272 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-nlvws" event={"ID":"46218023-25a1-49e3-8358-08445caf00ce","Type":"ContainerStarted","Data":"fb149d46dfe5eb769cecad148cab00998aa2fac8546e22b41cf3012056ebbb54"} Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.195110 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.196596 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlf9c\" (UniqueName: \"kubernetes.io/projected/d3d7ba45-9874-4177-81d0-9c8f88280cfe-kube-api-access-qlf9c\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.197098 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-config-data\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.197201 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-scripts\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.311250 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-config-data\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.311440 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-scripts\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.311624 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.311659 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlf9c\" (UniqueName: \"kubernetes.io/projected/d3d7ba45-9874-4177-81d0-9c8f88280cfe-kube-api-access-qlf9c\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.331195 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-scripts\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.358035 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.378541 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-config-data\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.384991 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-p7cb5" podStartSLOduration=4.384679379 podStartE2EDuration="4.384679379s" podCreationTimestamp="2026-03-17 16:06:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:06:13.139432234 +0000 UTC m=+1764.552748281" watchObservedRunningTime="2026-03-17 16:06:13.384679379 +0000 UTC m=+1764.797995426" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.392983 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlf9c\" (UniqueName: \"kubernetes.io/projected/d3d7ba45-9874-4177-81d0-9c8f88280cfe-kube-api-access-qlf9c\") pod \"nova-cell1-conductor-db-sync-9fv46\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.406282 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-nlvws" podStartSLOduration=5.406259406 podStartE2EDuration="5.406259406s" podCreationTimestamp="2026-03-17 16:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:06:13.175129895 +0000 UTC m=+1764.588445942" watchObservedRunningTime="2026-03-17 16:06:13.406259406 +0000 UTC m=+1764.819575453" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.462467 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.525825 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:06:13 crc kubenswrapper[4767]: I0317 16:06:13.549700 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 16:06:14 crc kubenswrapper[4767]: E0317 16:06:14.065017 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cf85c2_ceca_418b_a22e_ecadc2a3acfc.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:14 crc kubenswrapper[4767]: I0317 16:06:14.171606 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"64c09859-d250-40a0-97a1-aa8fd1371432","Type":"ContainerStarted","Data":"e9f71b08593019d707fdc106d2dba6494affea9ef12d9ae834fca6a59750f777"} Mar 17 16:06:14 crc kubenswrapper[4767]: I0317 16:06:14.174938 4767 generic.go:334] "Generic (PLEG): container finished" podID="46218023-25a1-49e3-8358-08445caf00ce" containerID="fb149d46dfe5eb769cecad148cab00998aa2fac8546e22b41cf3012056ebbb54" exitCode=0 Mar 17 16:06:14 crc kubenswrapper[4767]: I0317 16:06:14.174992 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-nlvws" event={"ID":"46218023-25a1-49e3-8358-08445caf00ce","Type":"ContainerDied","Data":"fb149d46dfe5eb769cecad148cab00998aa2fac8546e22b41cf3012056ebbb54"} Mar 17 16:06:14 crc kubenswrapper[4767]: I0317 16:06:14.182069 4767 generic.go:334] "Generic (PLEG): container finished" podID="271cc3d9-d20c-4708-bd6a-765a8c439cf6" containerID="eea4f2a1ea128dac6e5fa290ffcaf4f7e32b29d8e34f63c19f28c0eebd49d236" exitCode=0 Mar 17 16:06:14 crc kubenswrapper[4767]: I0317 16:06:14.182135 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-21c1-account-create-update-2g64g" event={"ID":"271cc3d9-d20c-4708-bd6a-765a8c439cf6","Type":"ContainerDied","Data":"eea4f2a1ea128dac6e5fa290ffcaf4f7e32b29d8e34f63c19f28c0eebd49d236"} Mar 17 16:06:14 crc kubenswrapper[4767]: I0317 16:06:14.184904 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fb28438-621e-41a8-b1ad-a07bf5e181c5","Type":"ContainerStarted","Data":"13c22b9e60f48132740d5d63e72b90ad413c9a23e241b579b5a241a12ec7bfd4"} Mar 17 16:06:14 crc kubenswrapper[4767]: I0317 16:06:14.187471 4767 generic.go:334] "Generic (PLEG): container finished" podID="f3467dd4-3ba8-4492-a678-611b6b982946" containerID="d33d736e32fc1fe05a1feabe53c8ce8244cb942ee434c5d90cfc10bf73e9ba89" exitCode=0 Mar 17 16:06:14 crc kubenswrapper[4767]: I0317 16:06:14.188165 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" event={"ID":"f3467dd4-3ba8-4492-a678-611b6b982946","Type":"ContainerDied","Data":"d33d736e32fc1fe05a1feabe53c8ce8244cb942ee434c5d90cfc10bf73e9ba89"} Mar 17 16:06:15 crc kubenswrapper[4767]: I0317 16:06:15.354950 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:06:15 crc kubenswrapper[4767]: E0317 16:06:15.355842 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:06:15 crc kubenswrapper[4767]: I0317 16:06:15.511944 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9fv46"] Mar 17 16:06:15 crc kubenswrapper[4767]: I0317 16:06:15.974340 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:06:16 crc kubenswrapper[4767]: I0317 16:06:16.055519 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 16:06:16 crc kubenswrapper[4767]: I0317 16:06:16.287088 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3586e791-e20d-44ea-a784-fd7c14389087","Type":"ContainerStarted","Data":"bd5e218de8595e00defc403afdf9a47d0e032674e683c6177fd0f6e38dc8aebe"} Mar 17 16:06:17 crc kubenswrapper[4767]: E0317 16:06:17.291705 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cf85c2_ceca_418b_a22e_ecadc2a3acfc.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.449108 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.637348 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-nlvws" Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.650964 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-21c1-account-create-update-2g64g" Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.682693 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7f88\" (UniqueName: \"kubernetes.io/projected/46218023-25a1-49e3-8358-08445caf00ce-kube-api-access-t7f88\") pod \"46218023-25a1-49e3-8358-08445caf00ce\" (UID: \"46218023-25a1-49e3-8358-08445caf00ce\") " Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.682778 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/271cc3d9-d20c-4708-bd6a-765a8c439cf6-operator-scripts\") pod \"271cc3d9-d20c-4708-bd6a-765a8c439cf6\" (UID: \"271cc3d9-d20c-4708-bd6a-765a8c439cf6\") " Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.682826 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46218023-25a1-49e3-8358-08445caf00ce-operator-scripts\") pod \"46218023-25a1-49e3-8358-08445caf00ce\" (UID: \"46218023-25a1-49e3-8358-08445caf00ce\") " Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.682929 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfbrh\" (UniqueName: \"kubernetes.io/projected/271cc3d9-d20c-4708-bd6a-765a8c439cf6-kube-api-access-mfbrh\") pod \"271cc3d9-d20c-4708-bd6a-765a8c439cf6\" (UID: \"271cc3d9-d20c-4708-bd6a-765a8c439cf6\") " Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.683453 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/271cc3d9-d20c-4708-bd6a-765a8c439cf6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "271cc3d9-d20c-4708-bd6a-765a8c439cf6" (UID: "271cc3d9-d20c-4708-bd6a-765a8c439cf6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.683831 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46218023-25a1-49e3-8358-08445caf00ce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "46218023-25a1-49e3-8358-08445caf00ce" (UID: "46218023-25a1-49e3-8358-08445caf00ce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.684252 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/271cc3d9-d20c-4708-bd6a-765a8c439cf6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.684275 4767 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46218023-25a1-49e3-8358-08445caf00ce-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.704221 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/271cc3d9-d20c-4708-bd6a-765a8c439cf6-kube-api-access-mfbrh" (OuterVolumeSpecName: "kube-api-access-mfbrh") pod "271cc3d9-d20c-4708-bd6a-765a8c439cf6" (UID: "271cc3d9-d20c-4708-bd6a-765a8c439cf6"). InnerVolumeSpecName "kube-api-access-mfbrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.704741 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46218023-25a1-49e3-8358-08445caf00ce-kube-api-access-t7f88" (OuterVolumeSpecName: "kube-api-access-t7f88") pod "46218023-25a1-49e3-8358-08445caf00ce" (UID: "46218023-25a1-49e3-8358-08445caf00ce"). InnerVolumeSpecName "kube-api-access-t7f88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.787340 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7f88\" (UniqueName: \"kubernetes.io/projected/46218023-25a1-49e3-8358-08445caf00ce-kube-api-access-t7f88\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:17 crc kubenswrapper[4767]: I0317 16:06:17.787379 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfbrh\" (UniqueName: \"kubernetes.io/projected/271cc3d9-d20c-4708-bd6a-765a8c439cf6-kube-api-access-mfbrh\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:18 crc kubenswrapper[4767]: I0317 16:06:18.504891 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-nlvws" event={"ID":"46218023-25a1-49e3-8358-08445caf00ce","Type":"ContainerDied","Data":"53e90e2cc857a3744cee05dad708820124c926eb163cd2b1320be31f59d979a8"} Mar 17 16:06:18 crc kubenswrapper[4767]: I0317 16:06:18.504950 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53e90e2cc857a3744cee05dad708820124c926eb163cd2b1320be31f59d979a8" Mar 17 16:06:18 crc kubenswrapper[4767]: I0317 16:06:18.505030 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-nlvws" Mar 17 16:06:18 crc kubenswrapper[4767]: I0317 16:06:18.512507 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-21c1-account-create-update-2g64g" event={"ID":"271cc3d9-d20c-4708-bd6a-765a8c439cf6","Type":"ContainerDied","Data":"dad31ae592e15fa8c1a179afcc2356f6d6a46f73f4bf1c15bd5d1afcf698a27a"} Mar 17 16:06:18 crc kubenswrapper[4767]: I0317 16:06:18.512550 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dad31ae592e15fa8c1a179afcc2356f6d6a46f73f4bf1c15bd5d1afcf698a27a" Mar 17 16:06:18 crc kubenswrapper[4767]: I0317 16:06:18.512622 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-21c1-account-create-update-2g64g" Mar 17 16:06:18 crc kubenswrapper[4767]: I0317 16:06:18.530838 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" event={"ID":"f3467dd4-3ba8-4492-a678-611b6b982946","Type":"ContainerStarted","Data":"f8b5447b2f04eaa526827541b47e97798418eb8585f7249cf7d8889d5722b52b"} Mar 17 16:06:18 crc kubenswrapper[4767]: I0317 16:06:18.531333 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:18 crc kubenswrapper[4767]: I0317 16:06:18.554754 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" podStartSLOduration=9.55473473 podStartE2EDuration="9.55473473s" podCreationTimestamp="2026-03-17 16:06:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:06:18.554248886 +0000 UTC m=+1769.967564943" watchObservedRunningTime="2026-03-17 16:06:18.55473473 +0000 UTC m=+1769.968050777" Mar 17 16:06:18 crc kubenswrapper[4767]: W0317 16:06:18.903381 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3d7ba45_9874_4177_81d0_9c8f88280cfe.slice/crio-3462f972728db6f625f1c4e0ac2a13e0b03db32ef3d857b076e1ff655bf786d7 WatchSource:0}: Error finding container 3462f972728db6f625f1c4e0ac2a13e0b03db32ef3d857b076e1ff655bf786d7: Status 404 returned error can't find the container with id 3462f972728db6f625f1c4e0ac2a13e0b03db32ef3d857b076e1ff655bf786d7 Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.587469 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9fv46" event={"ID":"d3d7ba45-9874-4177-81d0-9c8f88280cfe","Type":"ContainerStarted","Data":"3462f972728db6f625f1c4e0ac2a13e0b03db32ef3d857b076e1ff655bf786d7"} Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.926072 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-bj7s2"] Mar 17 16:06:19 crc kubenswrapper[4767]: E0317 16:06:19.926886 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46218023-25a1-49e3-8358-08445caf00ce" containerName="mariadb-database-create" Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.926913 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="46218023-25a1-49e3-8358-08445caf00ce" containerName="mariadb-database-create" Mar 17 16:06:19 crc kubenswrapper[4767]: E0317 16:06:19.926956 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271cc3d9-d20c-4708-bd6a-765a8c439cf6" containerName="mariadb-account-create-update" Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.926965 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="271cc3d9-d20c-4708-bd6a-765a8c439cf6" containerName="mariadb-account-create-update" Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.927309 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="271cc3d9-d20c-4708-bd6a-765a8c439cf6" containerName="mariadb-account-create-update" Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.927362 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="46218023-25a1-49e3-8358-08445caf00ce" containerName="mariadb-database-create" Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.928577 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.931256 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.931487 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.931482 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.932710 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-9wfxq" Mar 17 16:06:19 crc kubenswrapper[4767]: I0317 16:06:19.961335 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-bj7s2"] Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.117811 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcv8g\" (UniqueName: \"kubernetes.io/projected/9d52c8d8-4951-43d6-8889-60f379933fa7-kube-api-access-xcv8g\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.117880 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-config-data\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.117914 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-combined-ca-bundle\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.118478 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-scripts\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.224236 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcv8g\" (UniqueName: \"kubernetes.io/projected/9d52c8d8-4951-43d6-8889-60f379933fa7-kube-api-access-xcv8g\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.224580 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-config-data\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.225283 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-combined-ca-bundle\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.225862 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-scripts\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.233908 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-combined-ca-bundle\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.235022 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-scripts\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.236883 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-config-data\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.261016 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcv8g\" (UniqueName: \"kubernetes.io/projected/9d52c8d8-4951-43d6-8889-60f379933fa7-kube-api-access-xcv8g\") pod \"aodh-db-sync-bj7s2\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:20 crc kubenswrapper[4767]: I0317 16:06:20.268205 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.108699 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-bj7s2"] Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.629953 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"64c09859-d250-40a0-97a1-aa8fd1371432","Type":"ContainerStarted","Data":"76eed37fb5569df9487d3312c3195838a7de1e24fbe34ef038601b78a3adff01"} Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.630295 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="64c09859-d250-40a0-97a1-aa8fd1371432" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://76eed37fb5569df9487d3312c3195838a7de1e24fbe34ef038601b78a3adff01" gracePeriod=30 Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.635579 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3586e791-e20d-44ea-a784-fd7c14389087","Type":"ContainerStarted","Data":"bf214cef939e9d29079640483beaef088eabecc00bce545d152fb2510a16f13e"} Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.651859 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fb28438-621e-41a8-b1ad-a07bf5e181c5","Type":"ContainerStarted","Data":"361ad6c49e7669c253f3bc43711cc6f6a6b0ce56e8ef019a14e944705c3324eb"} Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.652272 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fb28438-621e-41a8-b1ad-a07bf5e181c5","Type":"ContainerStarted","Data":"9d8a897af0b9f6ed3b8a01486cdb6f477b0d3cb7be06b757dc35d75f02e26412"} Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.652594 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0fb28438-621e-41a8-b1ad-a07bf5e181c5" containerName="nova-metadata-log" containerID="cri-o://9d8a897af0b9f6ed3b8a01486cdb6f477b0d3cb7be06b757dc35d75f02e26412" gracePeriod=30 Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.653209 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0fb28438-621e-41a8-b1ad-a07bf5e181c5" containerName="nova-metadata-metadata" containerID="cri-o://361ad6c49e7669c253f3bc43711cc6f6a6b0ce56e8ef019a14e944705c3324eb" gracePeriod=30 Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.666134 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c60ded0-ef6b-4b07-93ef-b0aa6c643084","Type":"ContainerStarted","Data":"8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61"} Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.666471 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c60ded0-ef6b-4b07-93ef-b0aa6c643084","Type":"ContainerStarted","Data":"80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56"} Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.669332 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9fv46" event={"ID":"d3d7ba45-9874-4177-81d0-9c8f88280cfe","Type":"ContainerStarted","Data":"6de3fce5967ad59a9643a96e5e108e2b64cb299cb49ba3c0becfde2ed5e10f36"} Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.699452 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc","Type":"ContainerStarted","Data":"0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4"} Mar 17 16:06:21 crc kubenswrapper[4767]: I0317 16:06:21.716095 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-bj7s2" event={"ID":"9d52c8d8-4951-43d6-8889-60f379933fa7","Type":"ContainerStarted","Data":"618e466306a256e3d84fc500867a46cc357bc727db65e140f218130575ed22f6"} Mar 17 16:06:22 crc kubenswrapper[4767]: I0317 16:06:22.039458 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=5.139918181 podStartE2EDuration="12.039433528s" podCreationTimestamp="2026-03-17 16:06:10 +0000 UTC" firstStartedPulling="2026-03-17 16:06:13.442594915 +0000 UTC m=+1764.855910962" lastFinishedPulling="2026-03-17 16:06:20.342110262 +0000 UTC m=+1771.755426309" observedRunningTime="2026-03-17 16:06:21.9926474 +0000 UTC m=+1773.405963457" watchObservedRunningTime="2026-03-17 16:06:22.039433528 +0000 UTC m=+1773.452749575" Mar 17 16:06:22 crc kubenswrapper[4767]: I0317 16:06:22.041608 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=6.100941176 podStartE2EDuration="13.04159783s" podCreationTimestamp="2026-03-17 16:06:09 +0000 UTC" firstStartedPulling="2026-03-17 16:06:13.401391087 +0000 UTC m=+1764.814707124" lastFinishedPulling="2026-03-17 16:06:20.342047731 +0000 UTC m=+1771.755363778" observedRunningTime="2026-03-17 16:06:22.02308678 +0000 UTC m=+1773.436402847" watchObservedRunningTime="2026-03-17 16:06:22.04159783 +0000 UTC m=+1773.454913877" Mar 17 16:06:22 crc kubenswrapper[4767]: I0317 16:06:22.111752 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-9fv46" podStartSLOduration=10.111721105 podStartE2EDuration="10.111721105s" podCreationTimestamp="2026-03-17 16:06:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:06:22.05313888 +0000 UTC m=+1773.466454937" watchObservedRunningTime="2026-03-17 16:06:22.111721105 +0000 UTC m=+1773.525037172" Mar 17 16:06:22 crc kubenswrapper[4767]: I0317 16:06:22.272106 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=5.844637955 podStartE2EDuration="13.272071142s" podCreationTimestamp="2026-03-17 16:06:09 +0000 UTC" firstStartedPulling="2026-03-17 16:06:12.924444375 +0000 UTC m=+1764.337760422" lastFinishedPulling="2026-03-17 16:06:20.351877562 +0000 UTC m=+1771.765193609" observedRunningTime="2026-03-17 16:06:22.09929181 +0000 UTC m=+1773.512607867" watchObservedRunningTime="2026-03-17 16:06:22.272071142 +0000 UTC m=+1773.685387189" Mar 17 16:06:22 crc kubenswrapper[4767]: I0317 16:06:22.348490 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=5.5249165 podStartE2EDuration="13.348453466s" podCreationTimestamp="2026-03-17 16:06:09 +0000 UTC" firstStartedPulling="2026-03-17 16:06:12.518494294 +0000 UTC m=+1763.931810341" lastFinishedPulling="2026-03-17 16:06:20.34203126 +0000 UTC m=+1771.755347307" observedRunningTime="2026-03-17 16:06:22.129823153 +0000 UTC m=+1773.543139200" watchObservedRunningTime="2026-03-17 16:06:22.348453466 +0000 UTC m=+1773.761769513" Mar 17 16:06:22 crc kubenswrapper[4767]: I0317 16:06:22.759616 4767 generic.go:334] "Generic (PLEG): container finished" podID="0fb28438-621e-41a8-b1ad-a07bf5e181c5" containerID="9d8a897af0b9f6ed3b8a01486cdb6f477b0d3cb7be06b757dc35d75f02e26412" exitCode=143 Mar 17 16:06:22 crc kubenswrapper[4767]: I0317 16:06:22.759712 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fb28438-621e-41a8-b1ad-a07bf5e181c5","Type":"ContainerDied","Data":"9d8a897af0b9f6ed3b8a01486cdb6f477b0d3cb7be06b757dc35d75f02e26412"} Mar 17 16:06:24 crc kubenswrapper[4767]: I0317 16:06:24.801424 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3586e791-e20d-44ea-a784-fd7c14389087","Type":"ContainerStarted","Data":"2f5db0ab3c395fef62d0af21883615ddb7388dae6d2a0f5f9be5cb071ef8e67a"} Mar 17 16:06:24 crc kubenswrapper[4767]: I0317 16:06:24.801980 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="ceilometer-central-agent" containerID="cri-o://ec289db6fac44df87d03cf5f2520718277113a8a3b75ae87ec8a5c7ec99c7184" gracePeriod=30 Mar 17 16:06:24 crc kubenswrapper[4767]: I0317 16:06:24.802042 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 16:06:24 crc kubenswrapper[4767]: I0317 16:06:24.802116 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="sg-core" containerID="cri-o://bf214cef939e9d29079640483beaef088eabecc00bce545d152fb2510a16f13e" gracePeriod=30 Mar 17 16:06:24 crc kubenswrapper[4767]: I0317 16:06:24.802211 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="proxy-httpd" containerID="cri-o://2f5db0ab3c395fef62d0af21883615ddb7388dae6d2a0f5f9be5cb071ef8e67a" gracePeriod=30 Mar 17 16:06:24 crc kubenswrapper[4767]: I0317 16:06:24.802150 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="ceilometer-notification-agent" containerID="cri-o://bd5e218de8595e00defc403afdf9a47d0e032674e683c6177fd0f6e38dc8aebe" gracePeriod=30 Mar 17 16:06:24 crc kubenswrapper[4767]: I0317 16:06:24.858742 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.927694415 podStartE2EDuration="16.858715854s" podCreationTimestamp="2026-03-17 16:06:08 +0000 UTC" firstStartedPulling="2026-03-17 16:06:09.736080313 +0000 UTC m=+1761.149396360" lastFinishedPulling="2026-03-17 16:06:23.667101752 +0000 UTC m=+1775.080417799" observedRunningTime="2026-03-17 16:06:24.830208058 +0000 UTC m=+1776.243524105" watchObservedRunningTime="2026-03-17 16:06:24.858715854 +0000 UTC m=+1776.272031891" Mar 17 16:06:25 crc kubenswrapper[4767]: I0317 16:06:25.827284 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 17 16:06:25 crc kubenswrapper[4767]: I0317 16:06:25.833033 4767 generic.go:334] "Generic (PLEG): container finished" podID="3586e791-e20d-44ea-a784-fd7c14389087" containerID="2f5db0ab3c395fef62d0af21883615ddb7388dae6d2a0f5f9be5cb071ef8e67a" exitCode=0 Mar 17 16:06:25 crc kubenswrapper[4767]: I0317 16:06:25.833076 4767 generic.go:334] "Generic (PLEG): container finished" podID="3586e791-e20d-44ea-a784-fd7c14389087" containerID="bf214cef939e9d29079640483beaef088eabecc00bce545d152fb2510a16f13e" exitCode=2 Mar 17 16:06:25 crc kubenswrapper[4767]: I0317 16:06:25.833084 4767 generic.go:334] "Generic (PLEG): container finished" podID="3586e791-e20d-44ea-a784-fd7c14389087" containerID="bd5e218de8595e00defc403afdf9a47d0e032674e683c6177fd0f6e38dc8aebe" exitCode=0 Mar 17 16:06:25 crc kubenswrapper[4767]: I0317 16:06:25.833110 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3586e791-e20d-44ea-a784-fd7c14389087","Type":"ContainerDied","Data":"2f5db0ab3c395fef62d0af21883615ddb7388dae6d2a0f5f9be5cb071ef8e67a"} Mar 17 16:06:25 crc kubenswrapper[4767]: I0317 16:06:25.833143 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3586e791-e20d-44ea-a784-fd7c14389087","Type":"ContainerDied","Data":"bf214cef939e9d29079640483beaef088eabecc00bce545d152fb2510a16f13e"} Mar 17 16:06:25 crc kubenswrapper[4767]: I0317 16:06:25.833289 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3586e791-e20d-44ea-a784-fd7c14389087","Type":"ContainerDied","Data":"bd5e218de8595e00defc403afdf9a47d0e032674e683c6177fd0f6e38dc8aebe"} Mar 17 16:06:25 crc kubenswrapper[4767]: I0317 16:06:25.857483 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:06:25 crc kubenswrapper[4767]: I0317 16:06:25.946448 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:26 crc kubenswrapper[4767]: I0317 16:06:26.020950 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-q896r"] Mar 17 16:06:26 crc kubenswrapper[4767]: I0317 16:06:26.021606 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" podUID="87fb15fd-3cec-4d9a-a316-7f09d9a3062f" containerName="dnsmasq-dns" containerID="cri-o://87e9dbcc6a20ba9d0d5a91bfc60d63dbb651c0a0512804ea242ef42b5ac9c183" gracePeriod=10 Mar 17 16:06:26 crc kubenswrapper[4767]: I0317 16:06:26.850742 4767 generic.go:334] "Generic (PLEG): container finished" podID="87fb15fd-3cec-4d9a-a316-7f09d9a3062f" containerID="87e9dbcc6a20ba9d0d5a91bfc60d63dbb651c0a0512804ea242ef42b5ac9c183" exitCode=0 Mar 17 16:06:26 crc kubenswrapper[4767]: I0317 16:06:26.850844 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" event={"ID":"87fb15fd-3cec-4d9a-a316-7f09d9a3062f","Type":"ContainerDied","Data":"87e9dbcc6a20ba9d0d5a91bfc60d63dbb651c0a0512804ea242ef42b5ac9c183"} Mar 17 16:06:27 crc kubenswrapper[4767]: E0317 16:06:27.630770 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cf85c2_ceca_418b_a22e_ecadc2a3acfc.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:28 crc kubenswrapper[4767]: I0317 16:06:28.355640 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:06:28 crc kubenswrapper[4767]: E0317 16:06:28.358028 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:06:28 crc kubenswrapper[4767]: I0317 16:06:28.437612 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 17 16:06:28 crc kubenswrapper[4767]: I0317 16:06:28.437673 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 17 16:06:28 crc kubenswrapper[4767]: E0317 16:06:28.683943 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cf85c2_ceca_418b_a22e_ecadc2a3acfc.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:28 crc kubenswrapper[4767]: I0317 16:06:28.854845 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 17 16:06:28 crc kubenswrapper[4767]: I0317 16:06:28.855434 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 17 16:06:28 crc kubenswrapper[4767]: I0317 16:06:28.884097 4767 generic.go:334] "Generic (PLEG): container finished" podID="0aaae974-d49d-4980-a540-6bc2ec85a15f" containerID="466e5152b4dae52e57a86166113f0ef3d5745577ca45e1f223ca9a7f0065c3f8" exitCode=0 Mar 17 16:06:28 crc kubenswrapper[4767]: I0317 16:06:28.884215 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-p7cb5" event={"ID":"0aaae974-d49d-4980-a540-6bc2ec85a15f","Type":"ContainerDied","Data":"466e5152b4dae52e57a86166113f0ef3d5745577ca45e1f223ca9a7f0065c3f8"} Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.017438 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" podUID="87fb15fd-3cec-4d9a-a316-7f09d9a3062f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.223:5353: connect: connection refused" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.438018 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.438395 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.620459 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.676492 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-combined-ca-bundle\") pod \"0aaae974-d49d-4980-a540-6bc2ec85a15f\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.676636 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-config-data\") pod \"0aaae974-d49d-4980-a540-6bc2ec85a15f\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.676672 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cznlf\" (UniqueName: \"kubernetes.io/projected/0aaae974-d49d-4980-a540-6bc2ec85a15f-kube-api-access-cznlf\") pod \"0aaae974-d49d-4980-a540-6bc2ec85a15f\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.676878 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-scripts\") pod \"0aaae974-d49d-4980-a540-6bc2ec85a15f\" (UID: \"0aaae974-d49d-4980-a540-6bc2ec85a15f\") " Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.700806 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-scripts" (OuterVolumeSpecName: "scripts") pod "0aaae974-d49d-4980-a540-6bc2ec85a15f" (UID: "0aaae974-d49d-4980-a540-6bc2ec85a15f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.702456 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aaae974-d49d-4980-a540-6bc2ec85a15f-kube-api-access-cznlf" (OuterVolumeSpecName: "kube-api-access-cznlf") pod "0aaae974-d49d-4980-a540-6bc2ec85a15f" (UID: "0aaae974-d49d-4980-a540-6bc2ec85a15f"). InnerVolumeSpecName "kube-api-access-cznlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.763138 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0aaae974-d49d-4980-a540-6bc2ec85a15f" (UID: "0aaae974-d49d-4980-a540-6bc2ec85a15f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.763604 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-config-data" (OuterVolumeSpecName: "config-data") pod "0aaae974-d49d-4980-a540-6bc2ec85a15f" (UID: "0aaae974-d49d-4980-a540-6bc2ec85a15f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.781472 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.781518 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.781530 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cznlf\" (UniqueName: \"kubernetes.io/projected/0aaae974-d49d-4980-a540-6bc2ec85a15f-kube-api-access-cznlf\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.781547 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aaae974-d49d-4980-a540-6bc2ec85a15f-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.825500 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.835773 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.863874 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.955207 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" event={"ID":"87fb15fd-3cec-4d9a-a316-7f09d9a3062f","Type":"ContainerDied","Data":"0aa758e93aaf41290d8a31d89a468d95ae7aa16811530b8589d64f9fc539a306"} Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.955287 4767 scope.go:117] "RemoveContainer" containerID="87e9dbcc6a20ba9d0d5a91bfc60d63dbb651c0a0512804ea242ef42b5ac9c183" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.955512 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-q896r" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.966987 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-p7cb5" event={"ID":"0aaae974-d49d-4980-a540-6bc2ec85a15f","Type":"ContainerDied","Data":"973b7ace9ff4adc22087bd2be6b8d9340ebab4ddd6ce94fe8108f9f8bb7014ad"} Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.967039 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="973b7ace9ff4adc22087bd2be6b8d9340ebab4ddd6ce94fe8108f9f8bb7014ad" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.967123 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-p7cb5" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.976597 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-bj7s2" event={"ID":"9d52c8d8-4951-43d6-8889-60f379933fa7","Type":"ContainerStarted","Data":"52ab00d786ed6e40200bcdf22fa072d25c15f2ba3bec978aae02cda4cddbd0bb"} Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.993514 4767 scope.go:117] "RemoveContainer" containerID="b409be9c263934dcad6bec99c14f68c74de2cb7bba7ed96b1cd1dab4ca421d79" Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.997982 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-svc\") pod \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.998112 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbxsm\" (UniqueName: \"kubernetes.io/projected/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-kube-api-access-sbxsm\") pod \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.998394 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-nb\") pod \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.998611 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-config\") pod \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.998670 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-sb\") pod \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " Mar 17 16:06:30 crc kubenswrapper[4767]: I0317 16:06:30.998706 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-swift-storage-0\") pod \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\" (UID: \"87fb15fd-3cec-4d9a-a316-7f09d9a3062f\") " Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.006146 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-bj7s2" podStartSLOduration=2.754730045 podStartE2EDuration="12.005656007s" podCreationTimestamp="2026-03-17 16:06:19 +0000 UTC" firstStartedPulling="2026-03-17 16:06:21.106638588 +0000 UTC m=+1772.519954635" lastFinishedPulling="2026-03-17 16:06:30.35756455 +0000 UTC m=+1781.770880597" observedRunningTime="2026-03-17 16:06:30.995491256 +0000 UTC m=+1782.408807323" watchObservedRunningTime="2026-03-17 16:06:31.005656007 +0000 UTC m=+1782.418972064" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.017506 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-kube-api-access-sbxsm" (OuterVolumeSpecName: "kube-api-access-sbxsm") pod "87fb15fd-3cec-4d9a-a316-7f09d9a3062f" (UID: "87fb15fd-3cec-4d9a-a316-7f09d9a3062f"). InnerVolumeSpecName "kube-api-access-sbxsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.019619 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.102147 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-config" (OuterVolumeSpecName: "config") pod "87fb15fd-3cec-4d9a-a316-7f09d9a3062f" (UID: "87fb15fd-3cec-4d9a-a316-7f09d9a3062f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.102654 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.102823 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbxsm\" (UniqueName: \"kubernetes.io/projected/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-kube-api-access-sbxsm\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.110574 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "87fb15fd-3cec-4d9a-a316-7f09d9a3062f" (UID: "87fb15fd-3cec-4d9a-a316-7f09d9a3062f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.124932 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "87fb15fd-3cec-4d9a-a316-7f09d9a3062f" (UID: "87fb15fd-3cec-4d9a-a316-7f09d9a3062f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.151570 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "87fb15fd-3cec-4d9a-a316-7f09d9a3062f" (UID: "87fb15fd-3cec-4d9a-a316-7f09d9a3062f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.152810 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "87fb15fd-3cec-4d9a-a316-7f09d9a3062f" (UID: "87fb15fd-3cec-4d9a-a316-7f09d9a3062f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.214729 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.214779 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.214794 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.214805 4767 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87fb15fd-3cec-4d9a-a316-7f09d9a3062f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.219895 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.220332 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerName="nova-api-log" containerID="cri-o://80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56" gracePeriod=30 Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.220648 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerName="nova-api-api" containerID="cri-o://8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61" gracePeriod=30 Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.234988 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.250:8774/\": EOF" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.237682 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.250:8774/\": EOF" Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.525860 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-q896r"] Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.539865 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-q896r"] Mar 17 16:06:31 crc kubenswrapper[4767]: I0317 16:06:31.828480 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.118972 4767 generic.go:334] "Generic (PLEG): container finished" podID="d3d7ba45-9874-4177-81d0-9c8f88280cfe" containerID="6de3fce5967ad59a9643a96e5e108e2b64cb299cb49ba3c0becfde2ed5e10f36" exitCode=0 Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.119284 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9fv46" event={"ID":"d3d7ba45-9874-4177-81d0-9c8f88280cfe","Type":"ContainerDied","Data":"6de3fce5967ad59a9643a96e5e108e2b64cb299cb49ba3c0becfde2ed5e10f36"} Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.152361 4767 generic.go:334] "Generic (PLEG): container finished" podID="3586e791-e20d-44ea-a784-fd7c14389087" containerID="ec289db6fac44df87d03cf5f2520718277113a8a3b75ae87ec8a5c7ec99c7184" exitCode=0 Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.152464 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3586e791-e20d-44ea-a784-fd7c14389087","Type":"ContainerDied","Data":"ec289db6fac44df87d03cf5f2520718277113a8a3b75ae87ec8a5c7ec99c7184"} Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.174965 4767 generic.go:334] "Generic (PLEG): container finished" podID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerID="80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56" exitCode=143 Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.176474 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c60ded0-ef6b-4b07-93ef-b0aa6c643084","Type":"ContainerDied","Data":"80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56"} Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.671072 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.776832 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-run-httpd\") pod \"3586e791-e20d-44ea-a784-fd7c14389087\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.776940 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqmq9\" (UniqueName: \"kubernetes.io/projected/3586e791-e20d-44ea-a784-fd7c14389087-kube-api-access-xqmq9\") pod \"3586e791-e20d-44ea-a784-fd7c14389087\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.776976 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-sg-core-conf-yaml\") pod \"3586e791-e20d-44ea-a784-fd7c14389087\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.777211 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-scripts\") pod \"3586e791-e20d-44ea-a784-fd7c14389087\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.777283 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-log-httpd\") pod \"3586e791-e20d-44ea-a784-fd7c14389087\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.777384 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-combined-ca-bundle\") pod \"3586e791-e20d-44ea-a784-fd7c14389087\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.777526 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-config-data\") pod \"3586e791-e20d-44ea-a784-fd7c14389087\" (UID: \"3586e791-e20d-44ea-a784-fd7c14389087\") " Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.779569 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3586e791-e20d-44ea-a784-fd7c14389087" (UID: "3586e791-e20d-44ea-a784-fd7c14389087"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.783718 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3586e791-e20d-44ea-a784-fd7c14389087" (UID: "3586e791-e20d-44ea-a784-fd7c14389087"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.794616 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3586e791-e20d-44ea-a784-fd7c14389087-kube-api-access-xqmq9" (OuterVolumeSpecName: "kube-api-access-xqmq9") pod "3586e791-e20d-44ea-a784-fd7c14389087" (UID: "3586e791-e20d-44ea-a784-fd7c14389087"). InnerVolumeSpecName "kube-api-access-xqmq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.803054 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-scripts" (OuterVolumeSpecName: "scripts") pod "3586e791-e20d-44ea-a784-fd7c14389087" (UID: "3586e791-e20d-44ea-a784-fd7c14389087"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.880404 4767 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.880446 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqmq9\" (UniqueName: \"kubernetes.io/projected/3586e791-e20d-44ea-a784-fd7c14389087-kube-api-access-xqmq9\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.880492 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.880503 4767 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3586e791-e20d-44ea-a784-fd7c14389087-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.896519 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3586e791-e20d-44ea-a784-fd7c14389087" (UID: "3586e791-e20d-44ea-a784-fd7c14389087"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.967556 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3586e791-e20d-44ea-a784-fd7c14389087" (UID: "3586e791-e20d-44ea-a784-fd7c14389087"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.982486 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:32 crc kubenswrapper[4767]: I0317 16:06:32.982540 4767 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.020807 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-config-data" (OuterVolumeSpecName: "config-data") pod "3586e791-e20d-44ea-a784-fd7c14389087" (UID: "3586e791-e20d-44ea-a784-fd7c14389087"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.086129 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3586e791-e20d-44ea-a784-fd7c14389087-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.233329 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.252923 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3586e791-e20d-44ea-a784-fd7c14389087","Type":"ContainerDied","Data":"a7c2f8e0ca758482fa73a45c18b5ae75123121ff00219b9ffa0aad58960f77fa"} Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.252997 4767 scope.go:117] "RemoveContainer" containerID="2f5db0ab3c395fef62d0af21883615ddb7388dae6d2a0f5f9be5cb071ef8e67a" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.253322 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="dd2dafa6-8cc4-4f56-8155-d47b261a7bdc" containerName="nova-scheduler-scheduler" containerID="cri-o://0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4" gracePeriod=30 Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.334078 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.410027 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87fb15fd-3cec-4d9a-a316-7f09d9a3062f" path="/var/lib/kubelet/pods/87fb15fd-3cec-4d9a-a316-7f09d9a3062f/volumes" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.410826 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.410865 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:33 crc kubenswrapper[4767]: E0317 16:06:33.411383 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87fb15fd-3cec-4d9a-a316-7f09d9a3062f" containerName="init" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411405 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="87fb15fd-3cec-4d9a-a316-7f09d9a3062f" containerName="init" Mar 17 16:06:33 crc kubenswrapper[4767]: E0317 16:06:33.411430 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="proxy-httpd" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411440 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="proxy-httpd" Mar 17 16:06:33 crc kubenswrapper[4767]: E0317 16:06:33.411447 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="ceilometer-central-agent" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411454 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="ceilometer-central-agent" Mar 17 16:06:33 crc kubenswrapper[4767]: E0317 16:06:33.411478 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aaae974-d49d-4980-a540-6bc2ec85a15f" containerName="nova-manage" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411483 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aaae974-d49d-4980-a540-6bc2ec85a15f" containerName="nova-manage" Mar 17 16:06:33 crc kubenswrapper[4767]: E0317 16:06:33.411497 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87fb15fd-3cec-4d9a-a316-7f09d9a3062f" containerName="dnsmasq-dns" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411503 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="87fb15fd-3cec-4d9a-a316-7f09d9a3062f" containerName="dnsmasq-dns" Mar 17 16:06:33 crc kubenswrapper[4767]: E0317 16:06:33.411517 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="sg-core" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411523 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="sg-core" Mar 17 16:06:33 crc kubenswrapper[4767]: E0317 16:06:33.411537 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="ceilometer-notification-agent" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411543 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="ceilometer-notification-agent" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411788 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aaae974-d49d-4980-a540-6bc2ec85a15f" containerName="nova-manage" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411802 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="87fb15fd-3cec-4d9a-a316-7f09d9a3062f" containerName="dnsmasq-dns" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411813 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="ceilometer-notification-agent" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411826 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="proxy-httpd" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411845 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="ceilometer-central-agent" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.411855 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3586e791-e20d-44ea-a784-fd7c14389087" containerName="sg-core" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.413270 4767 scope.go:117] "RemoveContainer" containerID="bf214cef939e9d29079640483beaef088eabecc00bce545d152fb2510a16f13e" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.420853 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.425713 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.429820 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.433765 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.539451 4767 scope.go:117] "RemoveContainer" containerID="bd5e218de8595e00defc403afdf9a47d0e032674e683c6177fd0f6e38dc8aebe" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.616000 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-scripts\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.616101 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.616244 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdj92\" (UniqueName: \"kubernetes.io/projected/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-kube-api-access-hdj92\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.616296 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.616343 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-log-httpd\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.616392 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-config-data\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.616460 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-run-httpd\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.624562 4767 scope.go:117] "RemoveContainer" containerID="ec289db6fac44df87d03cf5f2520718277113a8a3b75ae87ec8a5c7ec99c7184" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.718429 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-log-httpd\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.718525 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-config-data\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.718597 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-run-httpd\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.718651 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-scripts\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.718694 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.718783 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdj92\" (UniqueName: \"kubernetes.io/projected/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-kube-api-access-hdj92\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.718829 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.724632 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-log-httpd\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.725782 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-run-httpd\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.739685 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-config-data\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.740436 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-scripts\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.748083 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.751248 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdj92\" (UniqueName: \"kubernetes.io/projected/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-kube-api-access-hdj92\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:33 crc kubenswrapper[4767]: I0317 16:06:33.764779 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " pod="openstack/ceilometer-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.011947 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.046203 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.132211 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlf9c\" (UniqueName: \"kubernetes.io/projected/d3d7ba45-9874-4177-81d0-9c8f88280cfe-kube-api-access-qlf9c\") pod \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.132301 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-combined-ca-bundle\") pod \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.132366 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-config-data\") pod \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.132407 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-scripts\") pod \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\" (UID: \"d3d7ba45-9874-4177-81d0-9c8f88280cfe\") " Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.139439 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3d7ba45-9874-4177-81d0-9c8f88280cfe-kube-api-access-qlf9c" (OuterVolumeSpecName: "kube-api-access-qlf9c") pod "d3d7ba45-9874-4177-81d0-9c8f88280cfe" (UID: "d3d7ba45-9874-4177-81d0-9c8f88280cfe"). InnerVolumeSpecName "kube-api-access-qlf9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.147338 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-scripts" (OuterVolumeSpecName: "scripts") pod "d3d7ba45-9874-4177-81d0-9c8f88280cfe" (UID: "d3d7ba45-9874-4177-81d0-9c8f88280cfe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.222813 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-config-data" (OuterVolumeSpecName: "config-data") pod "d3d7ba45-9874-4177-81d0-9c8f88280cfe" (UID: "d3d7ba45-9874-4177-81d0-9c8f88280cfe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.225546 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3d7ba45-9874-4177-81d0-9c8f88280cfe" (UID: "d3d7ba45-9874-4177-81d0-9c8f88280cfe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.246287 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.246334 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.246346 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3d7ba45-9874-4177-81d0-9c8f88280cfe-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.246357 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlf9c\" (UniqueName: \"kubernetes.io/projected/d3d7ba45-9874-4177-81d0-9c8f88280cfe-kube-api-access-qlf9c\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.303757 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9fv46" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.305388 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 17 16:06:34 crc kubenswrapper[4767]: E0317 16:06:34.306160 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3d7ba45-9874-4177-81d0-9c8f88280cfe" containerName="nova-cell1-conductor-db-sync" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.306204 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3d7ba45-9874-4177-81d0-9c8f88280cfe" containerName="nova-cell1-conductor-db-sync" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.306481 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3d7ba45-9874-4177-81d0-9c8f88280cfe" containerName="nova-cell1-conductor-db-sync" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.308340 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9fv46" event={"ID":"d3d7ba45-9874-4177-81d0-9c8f88280cfe","Type":"ContainerDied","Data":"3462f972728db6f625f1c4e0ac2a13e0b03db32ef3d857b076e1ff655bf786d7"} Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.308382 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3462f972728db6f625f1c4e0ac2a13e0b03db32ef3d857b076e1ff655bf786d7" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.308480 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.314518 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.373264 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.484812 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3709100-f2dc-4ae8-9051-51c01aa5a131-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d3709100-f2dc-4ae8-9051-51c01aa5a131\") " pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.485081 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3709100-f2dc-4ae8-9051-51c01aa5a131-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d3709100-f2dc-4ae8-9051-51c01aa5a131\") " pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.485218 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpcdh\" (UniqueName: \"kubernetes.io/projected/d3709100-f2dc-4ae8-9051-51c01aa5a131-kube-api-access-xpcdh\") pod \"nova-cell1-conductor-0\" (UID: \"d3709100-f2dc-4ae8-9051-51c01aa5a131\") " pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.588724 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3709100-f2dc-4ae8-9051-51c01aa5a131-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d3709100-f2dc-4ae8-9051-51c01aa5a131\") " pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.588800 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3709100-f2dc-4ae8-9051-51c01aa5a131-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d3709100-f2dc-4ae8-9051-51c01aa5a131\") " pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.588859 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpcdh\" (UniqueName: \"kubernetes.io/projected/d3709100-f2dc-4ae8-9051-51c01aa5a131-kube-api-access-xpcdh\") pod \"nova-cell1-conductor-0\" (UID: \"d3709100-f2dc-4ae8-9051-51c01aa5a131\") " pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.602159 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3709100-f2dc-4ae8-9051-51c01aa5a131-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d3709100-f2dc-4ae8-9051-51c01aa5a131\") " pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.603054 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3709100-f2dc-4ae8-9051-51c01aa5a131-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d3709100-f2dc-4ae8-9051-51c01aa5a131\") " pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.617283 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpcdh\" (UniqueName: \"kubernetes.io/projected/d3709100-f2dc-4ae8-9051-51c01aa5a131-kube-api-access-xpcdh\") pod \"nova-cell1-conductor-0\" (UID: \"d3709100-f2dc-4ae8-9051-51c01aa5a131\") " pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.668931 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:34 crc kubenswrapper[4767]: I0317 16:06:34.765004 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:34 crc kubenswrapper[4767]: W0317 16:06:34.769465 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2925ba9_6448_48b8_8ec0_3fb1830cfc01.slice/crio-8e6ba2ed89d9c00f5fff7cefa1ed8516341c0c335f9ea801071e33bf8fbfba0b WatchSource:0}: Error finding container 8e6ba2ed89d9c00f5fff7cefa1ed8516341c0c335f9ea801071e33bf8fbfba0b: Status 404 returned error can't find the container with id 8e6ba2ed89d9c00f5fff7cefa1ed8516341c0c335f9ea801071e33bf8fbfba0b Mar 17 16:06:35 crc kubenswrapper[4767]: I0317 16:06:35.237434 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 17 16:06:35 crc kubenswrapper[4767]: I0317 16:06:35.336519 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d3709100-f2dc-4ae8-9051-51c01aa5a131","Type":"ContainerStarted","Data":"3585dade31f3151517d496fb205467c18a232f4aa955bbc24eb28aa4c3a2b856"} Mar 17 16:06:35 crc kubenswrapper[4767]: I0317 16:06:35.338782 4767 generic.go:334] "Generic (PLEG): container finished" podID="9d52c8d8-4951-43d6-8889-60f379933fa7" containerID="52ab00d786ed6e40200bcdf22fa072d25c15f2ba3bec978aae02cda4cddbd0bb" exitCode=0 Mar 17 16:06:35 crc kubenswrapper[4767]: I0317 16:06:35.338896 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-bj7s2" event={"ID":"9d52c8d8-4951-43d6-8889-60f379933fa7","Type":"ContainerDied","Data":"52ab00d786ed6e40200bcdf22fa072d25c15f2ba3bec978aae02cda4cddbd0bb"} Mar 17 16:06:35 crc kubenswrapper[4767]: I0317 16:06:35.341784 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2925ba9-6448-48b8-8ec0-3fb1830cfc01","Type":"ContainerStarted","Data":"8e6ba2ed89d9c00f5fff7cefa1ed8516341c0c335f9ea801071e33bf8fbfba0b"} Mar 17 16:06:35 crc kubenswrapper[4767]: I0317 16:06:35.383995 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3586e791-e20d-44ea-a784-fd7c14389087" path="/var/lib/kubelet/pods/3586e791-e20d-44ea-a784-fd7c14389087/volumes" Mar 17 16:06:35 crc kubenswrapper[4767]: E0317 16:06:35.828099 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4 is running failed: container process not found" containerID="0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 17 16:06:35 crc kubenswrapper[4767]: E0317 16:06:35.828987 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4 is running failed: container process not found" containerID="0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 17 16:06:35 crc kubenswrapper[4767]: E0317 16:06:35.829307 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4 is running failed: container process not found" containerID="0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 17 16:06:35 crc kubenswrapper[4767]: E0317 16:06:35.829345 4767 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="dd2dafa6-8cc4-4f56-8155-d47b261a7bdc" containerName="nova-scheduler-scheduler" Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.423722 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d3709100-f2dc-4ae8-9051-51c01aa5a131","Type":"ContainerStarted","Data":"35fcecd60f11ad89e21aa2f005e61574845dc6cb3697cd09bb00cb8e5ad556f6"} Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.424136 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.429981 4767 generic.go:334] "Generic (PLEG): container finished" podID="dd2dafa6-8cc4-4f56-8155-d47b261a7bdc" containerID="0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4" exitCode=0 Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.430048 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc","Type":"ContainerDied","Data":"0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4"} Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.447846 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2925ba9-6448-48b8-8ec0-3fb1830cfc01","Type":"ContainerStarted","Data":"69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1"} Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.474107 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.474068911 podStartE2EDuration="2.474068911s" podCreationTimestamp="2026-03-17 16:06:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:06:36.459284848 +0000 UTC m=+1787.872600915" watchObservedRunningTime="2026-03-17 16:06:36.474068911 +0000 UTC m=+1787.887384968" Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.546192 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.710025 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-config-data\") pod \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.710223 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkjwf\" (UniqueName: \"kubernetes.io/projected/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-kube-api-access-wkjwf\") pod \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.710267 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-combined-ca-bundle\") pod \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\" (UID: \"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc\") " Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.746623 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-kube-api-access-wkjwf" (OuterVolumeSpecName: "kube-api-access-wkjwf") pod "dd2dafa6-8cc4-4f56-8155-d47b261a7bdc" (UID: "dd2dafa6-8cc4-4f56-8155-d47b261a7bdc"). InnerVolumeSpecName "kube-api-access-wkjwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.777356 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-config-data" (OuterVolumeSpecName: "config-data") pod "dd2dafa6-8cc4-4f56-8155-d47b261a7bdc" (UID: "dd2dafa6-8cc4-4f56-8155-d47b261a7bdc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.810335 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd2dafa6-8cc4-4f56-8155-d47b261a7bdc" (UID: "dd2dafa6-8cc4-4f56-8155-d47b261a7bdc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.816681 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkjwf\" (UniqueName: \"kubernetes.io/projected/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-kube-api-access-wkjwf\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.816729 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.816744 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:36 crc kubenswrapper[4767]: I0317 16:06:36.944486 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.023681 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-config-data\") pod \"9d52c8d8-4951-43d6-8889-60f379933fa7\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.023976 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcv8g\" (UniqueName: \"kubernetes.io/projected/9d52c8d8-4951-43d6-8889-60f379933fa7-kube-api-access-xcv8g\") pod \"9d52c8d8-4951-43d6-8889-60f379933fa7\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.024457 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-scripts\") pod \"9d52c8d8-4951-43d6-8889-60f379933fa7\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.024555 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-combined-ca-bundle\") pod \"9d52c8d8-4951-43d6-8889-60f379933fa7\" (UID: \"9d52c8d8-4951-43d6-8889-60f379933fa7\") " Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.030951 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-scripts" (OuterVolumeSpecName: "scripts") pod "9d52c8d8-4951-43d6-8889-60f379933fa7" (UID: "9d52c8d8-4951-43d6-8889-60f379933fa7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.038836 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d52c8d8-4951-43d6-8889-60f379933fa7-kube-api-access-xcv8g" (OuterVolumeSpecName: "kube-api-access-xcv8g") pod "9d52c8d8-4951-43d6-8889-60f379933fa7" (UID: "9d52c8d8-4951-43d6-8889-60f379933fa7"). InnerVolumeSpecName "kube-api-access-xcv8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.095684 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-config-data" (OuterVolumeSpecName: "config-data") pod "9d52c8d8-4951-43d6-8889-60f379933fa7" (UID: "9d52c8d8-4951-43d6-8889-60f379933fa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.102508 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d52c8d8-4951-43d6-8889-60f379933fa7" (UID: "9d52c8d8-4951-43d6-8889-60f379933fa7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.128914 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.128999 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcv8g\" (UniqueName: \"kubernetes.io/projected/9d52c8d8-4951-43d6-8889-60f379933fa7-kube-api-access-xcv8g\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.129012 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.129027 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d52c8d8-4951-43d6-8889-60f379933fa7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.466458 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-bj7s2" event={"ID":"9d52c8d8-4951-43d6-8889-60f379933fa7","Type":"ContainerDied","Data":"618e466306a256e3d84fc500867a46cc357bc727db65e140f218130575ed22f6"} Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.466826 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="618e466306a256e3d84fc500867a46cc357bc727db65e140f218130575ed22f6" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.466481 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-bj7s2" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.470009 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2925ba9-6448-48b8-8ec0-3fb1830cfc01","Type":"ContainerStarted","Data":"7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e"} Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.475332 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.476041 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dd2dafa6-8cc4-4f56-8155-d47b261a7bdc","Type":"ContainerDied","Data":"d3b6e2e96e0298b68e7e30bcee021aaa958092f32cd0e18fd0841fd07ba20124"} Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.476083 4767 scope.go:117] "RemoveContainer" containerID="0205e919f0b71c162b137fae1159a8634b964084de972b1829e0fbaa53b5e4c4" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.537901 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.560132 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.579302 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:06:37 crc kubenswrapper[4767]: E0317 16:06:37.580068 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d52c8d8-4951-43d6-8889-60f379933fa7" containerName="aodh-db-sync" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.580095 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d52c8d8-4951-43d6-8889-60f379933fa7" containerName="aodh-db-sync" Mar 17 16:06:37 crc kubenswrapper[4767]: E0317 16:06:37.580133 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2dafa6-8cc4-4f56-8155-d47b261a7bdc" containerName="nova-scheduler-scheduler" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.580140 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2dafa6-8cc4-4f56-8155-d47b261a7bdc" containerName="nova-scheduler-scheduler" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.580496 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d52c8d8-4951-43d6-8889-60f379933fa7" containerName="aodh-db-sync" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.580527 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd2dafa6-8cc4-4f56-8155-d47b261a7bdc" containerName="nova-scheduler-scheduler" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.581716 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.585586 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.592617 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.648957 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.649117 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qnxt\" (UniqueName: \"kubernetes.io/projected/2e690f7d-0a21-4296-b677-3c30f7cb878a-kube-api-access-6qnxt\") pod \"nova-scheduler-0\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.649277 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-config-data\") pod \"nova-scheduler-0\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.750646 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-config-data\") pod \"nova-scheduler-0\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.750780 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.750892 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qnxt\" (UniqueName: \"kubernetes.io/projected/2e690f7d-0a21-4296-b677-3c30f7cb878a-kube-api-access-6qnxt\") pod \"nova-scheduler-0\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.771417 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.772099 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-config-data\") pod \"nova-scheduler-0\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.795437 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qnxt\" (UniqueName: \"kubernetes.io/projected/2e690f7d-0a21-4296-b677-3c30f7cb878a-kube-api-access-6qnxt\") pod \"nova-scheduler-0\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: I0317 16:06:37.958685 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 16:06:37 crc kubenswrapper[4767]: E0317 16:06:37.991241 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cf85c2_ceca_418b_a22e_ecadc2a3acfc.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:38 crc kubenswrapper[4767]: I0317 16:06:38.489764 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2925ba9-6448-48b8-8ec0-3fb1830cfc01","Type":"ContainerStarted","Data":"c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5"} Mar 17 16:06:38 crc kubenswrapper[4767]: I0317 16:06:38.562733 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.134519 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.138416 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.141984 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-9wfxq" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.142316 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.144677 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.150714 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.303707 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.304421 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-scripts\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.304482 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-config-data\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.304565 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.304594 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnk8d\" (UniqueName: \"kubernetes.io/projected/f93b4e48-0fc9-4056-95c7-233caca8daf8-kube-api-access-wnk8d\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.377493 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:06:39 crc kubenswrapper[4767]: E0317 16:06:39.377839 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.398080 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd2dafa6-8cc4-4f56-8155-d47b261a7bdc" path="/var/lib/kubelet/pods/dd2dafa6-8cc4-4f56-8155-d47b261a7bdc/volumes" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.407319 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-config-data\") pod \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.407431 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-logs\") pod \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.407596 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw888\" (UniqueName: \"kubernetes.io/projected/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-kube-api-access-xw888\") pod \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.407684 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-combined-ca-bundle\") pod \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\" (UID: \"0c60ded0-ef6b-4b07-93ef-b0aa6c643084\") " Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.408250 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-scripts\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.408296 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-config-data\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.408345 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.408361 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnk8d\" (UniqueName: \"kubernetes.io/projected/f93b4e48-0fc9-4056-95c7-233caca8daf8-kube-api-access-wnk8d\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.414014 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-logs" (OuterVolumeSpecName: "logs") pod "0c60ded0-ef6b-4b07-93ef-b0aa6c643084" (UID: "0c60ded0-ef6b-4b07-93ef-b0aa6c643084"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.439215 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.444206 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-scripts\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.451051 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-kube-api-access-xw888" (OuterVolumeSpecName: "kube-api-access-xw888") pod "0c60ded0-ef6b-4b07-93ef-b0aa6c643084" (UID: "0c60ded0-ef6b-4b07-93ef-b0aa6c643084"). InnerVolumeSpecName "kube-api-access-xw888". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.470030 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnk8d\" (UniqueName: \"kubernetes.io/projected/f93b4e48-0fc9-4056-95c7-233caca8daf8-kube-api-access-wnk8d\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.473886 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c60ded0-ef6b-4b07-93ef-b0aa6c643084" (UID: "0c60ded0-ef6b-4b07-93ef-b0aa6c643084"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.487253 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-config-data" (OuterVolumeSpecName: "config-data") pod "0c60ded0-ef6b-4b07-93ef-b0aa6c643084" (UID: "0c60ded0-ef6b-4b07-93ef-b0aa6c643084"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.489745 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-config-data\") pod \"aodh-0\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.512721 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw888\" (UniqueName: \"kubernetes.io/projected/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-kube-api-access-xw888\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.512760 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.512772 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.512783 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c60ded0-ef6b-4b07-93ef-b0aa6c643084-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.553762 4767 generic.go:334] "Generic (PLEG): container finished" podID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerID="8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61" exitCode=0 Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.553878 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c60ded0-ef6b-4b07-93ef-b0aa6c643084","Type":"ContainerDied","Data":"8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61"} Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.553913 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0c60ded0-ef6b-4b07-93ef-b0aa6c643084","Type":"ContainerDied","Data":"6f3a64a53fad6d3b105ce6775ffee382fe9512563d49e10b7e85eee2429ce596"} Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.553963 4767 scope.go:117] "RemoveContainer" containerID="8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.554144 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.570667 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e690f7d-0a21-4296-b677-3c30f7cb878a","Type":"ContainerStarted","Data":"dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926"} Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.570989 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e690f7d-0a21-4296-b677-3c30f7cb878a","Type":"ContainerStarted","Data":"b6c47559be33d46a654ed9df5f211bcfb4c6bec895532c836357ab58cfd3ae1e"} Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.635506 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.635474142 podStartE2EDuration="2.635474142s" podCreationTimestamp="2026-03-17 16:06:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:06:39.596384154 +0000 UTC m=+1791.009700231" watchObservedRunningTime="2026-03-17 16:06:39.635474142 +0000 UTC m=+1791.048790189" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.690920 4767 scope.go:117] "RemoveContainer" containerID="80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.700045 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.719752 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.776682 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.794908 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 17 16:06:39 crc kubenswrapper[4767]: E0317 16:06:39.801401 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerName="nova-api-api" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.801453 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerName="nova-api-api" Mar 17 16:06:39 crc kubenswrapper[4767]: E0317 16:06:39.801506 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerName="nova-api-log" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.801514 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerName="nova-api-log" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.805448 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerName="nova-api-log" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.805502 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" containerName="nova-api-api" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.845028 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.845195 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.845315 4767 scope.go:117] "RemoveContainer" containerID="8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.850156 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 17 16:06:39 crc kubenswrapper[4767]: E0317 16:06:39.851598 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61\": container with ID starting with 8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61 not found: ID does not exist" containerID="8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.851642 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61"} err="failed to get container status \"8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61\": rpc error: code = NotFound desc = could not find container \"8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61\": container with ID starting with 8d3a99a92c4ac37f158e59e9fd0bc00e30ce7319c8f8ed411cbd49cc9a2e6c61 not found: ID does not exist" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.851676 4767 scope.go:117] "RemoveContainer" containerID="80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56" Mar 17 16:06:39 crc kubenswrapper[4767]: E0317 16:06:39.852188 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56\": container with ID starting with 80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56 not found: ID does not exist" containerID="80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56" Mar 17 16:06:39 crc kubenswrapper[4767]: I0317 16:06:39.852217 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56"} err="failed to get container status \"80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56\": rpc error: code = NotFound desc = could not find container \"80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56\": container with ID starting with 80e634cae35159a56f7cb2583c165c499c234714ebd64c529bbb5b8f979ebb56 not found: ID does not exist" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.045879 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.045934 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc5lq\" (UniqueName: \"kubernetes.io/projected/85921833-d8a5-4fdf-8687-5607b22b263b-kube-api-access-tc5lq\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.046089 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-config-data\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.046209 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85921833-d8a5-4fdf-8687-5607b22b263b-logs\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.313204 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.313263 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc5lq\" (UniqueName: \"kubernetes.io/projected/85921833-d8a5-4fdf-8687-5607b22b263b-kube-api-access-tc5lq\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.313528 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-config-data\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.313691 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85921833-d8a5-4fdf-8687-5607b22b263b-logs\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.316040 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85921833-d8a5-4fdf-8687-5607b22b263b-logs\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.331426 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.342261 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-config-data\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.365065 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc5lq\" (UniqueName: \"kubernetes.io/projected/85921833-d8a5-4fdf-8687-5607b22b263b-kube-api-access-tc5lq\") pod \"nova-api-0\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.635521 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.658837 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2925ba9-6448-48b8-8ec0-3fb1830cfc01","Type":"ContainerStarted","Data":"937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745"} Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.660035 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.714806 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.014486168 podStartE2EDuration="7.714781913s" podCreationTimestamp="2026-03-17 16:06:33 +0000 UTC" firstStartedPulling="2026-03-17 16:06:34.774725758 +0000 UTC m=+1786.188041805" lastFinishedPulling="2026-03-17 16:06:39.475021503 +0000 UTC m=+1790.888337550" observedRunningTime="2026-03-17 16:06:40.712903319 +0000 UTC m=+1792.126219366" watchObservedRunningTime="2026-03-17 16:06:40.714781913 +0000 UTC m=+1792.128097960" Mar 17 16:06:40 crc kubenswrapper[4767]: I0317 16:06:40.842388 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 17 16:06:41 crc kubenswrapper[4767]: I0317 16:06:41.371107 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c60ded0-ef6b-4b07-93ef-b0aa6c643084" path="/var/lib/kubelet/pods/0c60ded0-ef6b-4b07-93ef-b0aa6c643084/volumes" Mar 17 16:06:41 crc kubenswrapper[4767]: I0317 16:06:41.445227 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:06:41 crc kubenswrapper[4767]: W0317 16:06:41.470416 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85921833_d8a5_4fdf_8687_5607b22b263b.slice/crio-f67473e2e8fc943cc67c322282a752054e4318da10b1d0037a307255d1f4e03a WatchSource:0}: Error finding container f67473e2e8fc943cc67c322282a752054e4318da10b1d0037a307255d1f4e03a: Status 404 returned error can't find the container with id f67473e2e8fc943cc67c322282a752054e4318da10b1d0037a307255d1f4e03a Mar 17 16:06:41 crc kubenswrapper[4767]: I0317 16:06:41.679959 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85921833-d8a5-4fdf-8687-5607b22b263b","Type":"ContainerStarted","Data":"f67473e2e8fc943cc67c322282a752054e4318da10b1d0037a307255d1f4e03a"} Mar 17 16:06:41 crc kubenswrapper[4767]: I0317 16:06:41.686692 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"f93b4e48-0fc9-4056-95c7-233caca8daf8","Type":"ContainerStarted","Data":"4f2bee65bec5fc21fef6a2036ea4a22d107fc422dcb79b0b1adcf52a58540c7f"} Mar 17 16:06:42 crc kubenswrapper[4767]: I0317 16:06:42.703928 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85921833-d8a5-4fdf-8687-5607b22b263b","Type":"ContainerStarted","Data":"a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8"} Mar 17 16:06:42 crc kubenswrapper[4767]: I0317 16:06:42.705321 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85921833-d8a5-4fdf-8687-5607b22b263b","Type":"ContainerStarted","Data":"101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666"} Mar 17 16:06:42 crc kubenswrapper[4767]: I0317 16:06:42.958730 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 17 16:06:43 crc kubenswrapper[4767]: I0317 16:06:43.316160 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.316136635 podStartE2EDuration="4.316136635s" podCreationTimestamp="2026-03-17 16:06:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:06:42.735281422 +0000 UTC m=+1794.148597479" watchObservedRunningTime="2026-03-17 16:06:43.316136635 +0000 UTC m=+1794.729452682" Mar 17 16:06:43 crc kubenswrapper[4767]: I0317 16:06:43.331014 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Mar 17 16:06:43 crc kubenswrapper[4767]: I0317 16:06:43.904989 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"f93b4e48-0fc9-4056-95c7-233caca8daf8","Type":"ContainerStarted","Data":"a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280"} Mar 17 16:06:44 crc kubenswrapper[4767]: E0317 16:06:44.155607 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cf85c2_ceca_418b_a22e_ecadc2a3acfc.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:44 crc kubenswrapper[4767]: I0317 16:06:44.720152 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 17 16:06:45 crc kubenswrapper[4767]: I0317 16:06:45.938097 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"f93b4e48-0fc9-4056-95c7-233caca8daf8","Type":"ContainerStarted","Data":"839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3"} Mar 17 16:06:46 crc kubenswrapper[4767]: I0317 16:06:46.315645 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:46 crc kubenswrapper[4767]: I0317 16:06:46.316313 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="ceilometer-central-agent" containerID="cri-o://69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1" gracePeriod=30 Mar 17 16:06:46 crc kubenswrapper[4767]: I0317 16:06:46.316355 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="proxy-httpd" containerID="cri-o://937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745" gracePeriod=30 Mar 17 16:06:46 crc kubenswrapper[4767]: I0317 16:06:46.316394 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="ceilometer-notification-agent" containerID="cri-o://7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e" gracePeriod=30 Mar 17 16:06:46 crc kubenswrapper[4767]: I0317 16:06:46.316429 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="sg-core" containerID="cri-o://c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5" gracePeriod=30 Mar 17 16:06:47 crc kubenswrapper[4767]: I0317 16:06:47.014800 4767 generic.go:334] "Generic (PLEG): container finished" podID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerID="937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745" exitCode=0 Mar 17 16:06:47 crc kubenswrapper[4767]: I0317 16:06:47.014855 4767 generic.go:334] "Generic (PLEG): container finished" podID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerID="c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5" exitCode=2 Mar 17 16:06:47 crc kubenswrapper[4767]: I0317 16:06:47.014893 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2925ba9-6448-48b8-8ec0-3fb1830cfc01","Type":"ContainerDied","Data":"937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745"} Mar 17 16:06:47 crc kubenswrapper[4767]: I0317 16:06:47.014941 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2925ba9-6448-48b8-8ec0-3fb1830cfc01","Type":"ContainerDied","Data":"c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5"} Mar 17 16:06:47 crc kubenswrapper[4767]: I0317 16:06:47.959707 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 17 16:06:48 crc kubenswrapper[4767]: I0317 16:06:48.001967 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 17 16:06:48 crc kubenswrapper[4767]: E0317 16:06:48.050407 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cf85c2_ceca_418b_a22e_ecadc2a3acfc.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:48 crc kubenswrapper[4767]: I0317 16:06:48.055628 4767 generic.go:334] "Generic (PLEG): container finished" podID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerID="7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e" exitCode=0 Mar 17 16:06:48 crc kubenswrapper[4767]: I0317 16:06:48.056221 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2925ba9-6448-48b8-8ec0-3fb1830cfc01","Type":"ContainerDied","Data":"7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e"} Mar 17 16:06:48 crc kubenswrapper[4767]: I0317 16:06:48.096994 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 17 16:06:48 crc kubenswrapper[4767]: E0317 16:06:48.106326 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cf85c2_ceca_418b_a22e_ecadc2a3acfc.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:48 crc kubenswrapper[4767]: E0317 16:06:48.115314 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cf85c2_ceca_418b_a22e_ecadc2a3acfc.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:06:49 crc kubenswrapper[4767]: I0317 16:06:49.080496 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"f93b4e48-0fc9-4056-95c7-233caca8daf8","Type":"ContainerStarted","Data":"d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd"} Mar 17 16:06:50 crc kubenswrapper[4767]: I0317 16:06:50.638246 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 16:06:50 crc kubenswrapper[4767]: I0317 16:06:50.638740 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.106336 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.153252 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"f93b4e48-0fc9-4056-95c7-233caca8daf8","Type":"ContainerStarted","Data":"109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb"} Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.153948 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-api" containerID="cri-o://a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280" gracePeriod=30 Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.153981 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-evaluator" containerID="cri-o://839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3" gracePeriod=30 Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.154012 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-notifier" containerID="cri-o://d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd" gracePeriod=30 Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.154048 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-listener" containerID="cri-o://109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb" gracePeriod=30 Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.180008 4767 generic.go:334] "Generic (PLEG): container finished" podID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerID="69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1" exitCode=0 Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.180063 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2925ba9-6448-48b8-8ec0-3fb1830cfc01","Type":"ContainerDied","Data":"69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1"} Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.180097 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2925ba9-6448-48b8-8ec0-3fb1830cfc01","Type":"ContainerDied","Data":"8e6ba2ed89d9c00f5fff7cefa1ed8516341c0c335f9ea801071e33bf8fbfba0b"} Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.180117 4767 scope.go:117] "RemoveContainer" containerID="937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.180478 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.218289 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.477749565 podStartE2EDuration="12.218261949s" podCreationTimestamp="2026-03-17 16:06:39 +0000 UTC" firstStartedPulling="2026-03-17 16:06:40.868398556 +0000 UTC m=+1792.281714603" lastFinishedPulling="2026-03-17 16:06:50.60891094 +0000 UTC m=+1802.022226987" observedRunningTime="2026-03-17 16:06:51.195351733 +0000 UTC m=+1802.608667780" watchObservedRunningTime="2026-03-17 16:06:51.218261949 +0000 UTC m=+1802.631577996" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.240319 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-config-data\") pod \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.240495 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-run-httpd\") pod \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.240555 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdj92\" (UniqueName: \"kubernetes.io/projected/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-kube-api-access-hdj92\") pod \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.240610 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-combined-ca-bundle\") pod \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.240728 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-log-httpd\") pod \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.240895 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-sg-core-conf-yaml\") pod \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.240923 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-scripts\") pod \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\" (UID: \"d2925ba9-6448-48b8-8ec0-3fb1830cfc01\") " Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.241268 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d2925ba9-6448-48b8-8ec0-3fb1830cfc01" (UID: "d2925ba9-6448-48b8-8ec0-3fb1830cfc01"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.241602 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d2925ba9-6448-48b8-8ec0-3fb1830cfc01" (UID: "d2925ba9-6448-48b8-8ec0-3fb1830cfc01"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.241737 4767 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.253515 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-kube-api-access-hdj92" (OuterVolumeSpecName: "kube-api-access-hdj92") pod "d2925ba9-6448-48b8-8ec0-3fb1830cfc01" (UID: "d2925ba9-6448-48b8-8ec0-3fb1830cfc01"). InnerVolumeSpecName "kube-api-access-hdj92". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.260529 4767 scope.go:117] "RemoveContainer" containerID="c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.267764 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-scripts" (OuterVolumeSpecName: "scripts") pod "d2925ba9-6448-48b8-8ec0-3fb1830cfc01" (UID: "d2925ba9-6448-48b8-8ec0-3fb1830cfc01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.344670 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdj92\" (UniqueName: \"kubernetes.io/projected/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-kube-api-access-hdj92\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.345816 4767 scope.go:117] "RemoveContainer" containerID="7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.353095 4767 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.353131 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.375972 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d2925ba9-6448-48b8-8ec0-3fb1830cfc01" (UID: "d2925ba9-6448-48b8-8ec0-3fb1830cfc01"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.437437 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2925ba9-6448-48b8-8ec0-3fb1830cfc01" (UID: "d2925ba9-6448-48b8-8ec0-3fb1830cfc01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.454274 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-config-data" (OuterVolumeSpecName: "config-data") pod "d2925ba9-6448-48b8-8ec0-3fb1830cfc01" (UID: "d2925ba9-6448-48b8-8ec0-3fb1830cfc01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.455609 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.455649 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.455667 4767 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2925ba9-6448-48b8-8ec0-3fb1830cfc01-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.491400 4767 scope.go:117] "RemoveContainer" containerID="69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.564208 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.581896 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.596776 4767 scope.go:117] "RemoveContainer" containerID="937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745" Mar 17 16:06:51 crc kubenswrapper[4767]: E0317 16:06:51.600321 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745\": container with ID starting with 937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745 not found: ID does not exist" containerID="937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.600381 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745"} err="failed to get container status \"937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745\": rpc error: code = NotFound desc = could not find container \"937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745\": container with ID starting with 937da8bcc8bef7f9a82cc098e9ec298fce17394e691ed6d899777bba8a84c745 not found: ID does not exist" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.600410 4767 scope.go:117] "RemoveContainer" containerID="c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5" Mar 17 16:06:51 crc kubenswrapper[4767]: E0317 16:06:51.604040 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5\": container with ID starting with c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5 not found: ID does not exist" containerID="c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.604099 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5"} err="failed to get container status \"c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5\": rpc error: code = NotFound desc = could not find container \"c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5\": container with ID starting with c4481c2cc1c6e9d4398f2a9069f39d716360bbff4d65f4f7f8110f4a269538b5 not found: ID does not exist" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.604142 4767 scope.go:117] "RemoveContainer" containerID="7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e" Mar 17 16:06:51 crc kubenswrapper[4767]: E0317 16:06:51.608893 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e\": container with ID starting with 7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e not found: ID does not exist" containerID="7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.608933 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e"} err="failed to get container status \"7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e\": rpc error: code = NotFound desc = could not find container \"7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e\": container with ID starting with 7295a6c1b3c216c279718b824ac2e4943e0e5f6cfb389f2bf1a680cd0a44135e not found: ID does not exist" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.608960 4767 scope.go:117] "RemoveContainer" containerID="69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1" Mar 17 16:06:51 crc kubenswrapper[4767]: E0317 16:06:51.613291 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1\": container with ID starting with 69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1 not found: ID does not exist" containerID="69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.613343 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1"} err="failed to get container status \"69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1\": rpc error: code = NotFound desc = could not find container \"69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1\": container with ID starting with 69d6bd1c5dfcef0347576aa620d098cfa41516880963f8a831d766b09f08a2c1 not found: ID does not exist" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.643326 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:51 crc kubenswrapper[4767]: E0317 16:06:51.644104 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="ceilometer-notification-agent" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.644120 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="ceilometer-notification-agent" Mar 17 16:06:51 crc kubenswrapper[4767]: E0317 16:06:51.644139 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="proxy-httpd" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.644145 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="proxy-httpd" Mar 17 16:06:51 crc kubenswrapper[4767]: E0317 16:06:51.644152 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="sg-core" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.644159 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="sg-core" Mar 17 16:06:51 crc kubenswrapper[4767]: E0317 16:06:51.644237 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="ceilometer-central-agent" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.644244 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="ceilometer-central-agent" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.644583 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="ceilometer-central-agent" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.644647 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="ceilometer-notification-agent" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.644664 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="sg-core" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.644672 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" containerName="proxy-httpd" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.667072 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.667644 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.672330 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.673396 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.727833 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="85921833-d8a5-4fdf-8687-5607b22b263b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.5:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.728236 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="85921833-d8a5-4fdf-8687-5607b22b263b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.5:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.767456 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-scripts\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.767592 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-run-httpd\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.767683 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kr92\" (UniqueName: \"kubernetes.io/projected/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-kube-api-access-9kr92\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.767825 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.767882 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-config-data\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.767956 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-log-httpd\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.768338 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.870808 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-log-httpd\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.871017 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.871497 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-log-httpd\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.872357 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-scripts\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.872443 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-run-httpd\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.872479 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kr92\" (UniqueName: \"kubernetes.io/projected/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-kube-api-access-9kr92\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.872553 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.872614 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-config-data\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.873191 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-run-httpd\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.886008 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.887117 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-scripts\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.887567 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-config-data\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.910895 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:51 crc kubenswrapper[4767]: I0317 16:06:51.924190 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kr92\" (UniqueName: \"kubernetes.io/projected/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-kube-api-access-9kr92\") pod \"ceilometer-0\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " pod="openstack/ceilometer-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.069693 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.225011 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.226331 4767 generic.go:334] "Generic (PLEG): container finished" podID="0fb28438-621e-41a8-b1ad-a07bf5e181c5" containerID="361ad6c49e7669c253f3bc43711cc6f6a6b0ce56e8ef019a14e944705c3324eb" exitCode=137 Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.226390 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fb28438-621e-41a8-b1ad-a07bf5e181c5","Type":"ContainerDied","Data":"361ad6c49e7669c253f3bc43711cc6f6a6b0ce56e8ef019a14e944705c3324eb"} Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.231716 4767 generic.go:334] "Generic (PLEG): container finished" podID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerID="d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd" exitCode=0 Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.231745 4767 generic.go:334] "Generic (PLEG): container finished" podID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerID="839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3" exitCode=0 Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.231752 4767 generic.go:334] "Generic (PLEG): container finished" podID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerID="a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280" exitCode=0 Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.231832 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"f93b4e48-0fc9-4056-95c7-233caca8daf8","Type":"ContainerDied","Data":"d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd"} Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.231864 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"f93b4e48-0fc9-4056-95c7-233caca8daf8","Type":"ContainerDied","Data":"839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3"} Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.231875 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"f93b4e48-0fc9-4056-95c7-233caca8daf8","Type":"ContainerDied","Data":"a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280"} Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.236759 4767 generic.go:334] "Generic (PLEG): container finished" podID="64c09859-d250-40a0-97a1-aa8fd1371432" containerID="76eed37fb5569df9487d3312c3195838a7de1e24fbe34ef038601b78a3adff01" exitCode=137 Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.236812 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"64c09859-d250-40a0-97a1-aa8fd1371432","Type":"ContainerDied","Data":"76eed37fb5569df9487d3312c3195838a7de1e24fbe34ef038601b78a3adff01"} Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.236852 4767 scope.go:117] "RemoveContainer" containerID="76eed37fb5569df9487d3312c3195838a7de1e24fbe34ef038601b78a3adff01" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.237034 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.391681 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-config-data\") pod \"64c09859-d250-40a0-97a1-aa8fd1371432\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.392447 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8sz2\" (UniqueName: \"kubernetes.io/projected/64c09859-d250-40a0-97a1-aa8fd1371432-kube-api-access-z8sz2\") pod \"64c09859-d250-40a0-97a1-aa8fd1371432\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.392809 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-combined-ca-bundle\") pod \"64c09859-d250-40a0-97a1-aa8fd1371432\" (UID: \"64c09859-d250-40a0-97a1-aa8fd1371432\") " Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.406153 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64c09859-d250-40a0-97a1-aa8fd1371432-kube-api-access-z8sz2" (OuterVolumeSpecName: "kube-api-access-z8sz2") pod "64c09859-d250-40a0-97a1-aa8fd1371432" (UID: "64c09859-d250-40a0-97a1-aa8fd1371432"). InnerVolumeSpecName "kube-api-access-z8sz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.446568 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-config-data" (OuterVolumeSpecName: "config-data") pod "64c09859-d250-40a0-97a1-aa8fd1371432" (UID: "64c09859-d250-40a0-97a1-aa8fd1371432"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.460230 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64c09859-d250-40a0-97a1-aa8fd1371432" (UID: "64c09859-d250-40a0-97a1-aa8fd1371432"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.510293 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8sz2\" (UniqueName: \"kubernetes.io/projected/64c09859-d250-40a0-97a1-aa8fd1371432-kube-api-access-z8sz2\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.510343 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.510354 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c09859-d250-40a0-97a1-aa8fd1371432-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.515932 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.623761 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb28438-621e-41a8-b1ad-a07bf5e181c5-logs\") pod \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.624134 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-config-data\") pod \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.624300 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-combined-ca-bundle\") pod \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.626160 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fb28438-621e-41a8-b1ad-a07bf5e181c5-logs" (OuterVolumeSpecName: "logs") pod "0fb28438-621e-41a8-b1ad-a07bf5e181c5" (UID: "0fb28438-621e-41a8-b1ad-a07bf5e181c5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.628832 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb28438-621e-41a8-b1ad-a07bf5e181c5-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.677463 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fb28438-621e-41a8-b1ad-a07bf5e181c5" (UID: "0fb28438-621e-41a8-b1ad-a07bf5e181c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.721885 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.730509 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjwk2\" (UniqueName: \"kubernetes.io/projected/0fb28438-621e-41a8-b1ad-a07bf5e181c5-kube-api-access-cjwk2\") pod \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\" (UID: \"0fb28438-621e-41a8-b1ad-a07bf5e181c5\") " Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.731741 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.750579 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fb28438-621e-41a8-b1ad-a07bf5e181c5-kube-api-access-cjwk2" (OuterVolumeSpecName: "kube-api-access-cjwk2") pod "0fb28438-621e-41a8-b1ad-a07bf5e181c5" (UID: "0fb28438-621e-41a8-b1ad-a07bf5e181c5"). InnerVolumeSpecName "kube-api-access-cjwk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.779440 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.784684 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-config-data" (OuterVolumeSpecName: "config-data") pod "0fb28438-621e-41a8-b1ad-a07bf5e181c5" (UID: "0fb28438-621e-41a8-b1ad-a07bf5e181c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.815309 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 16:06:52 crc kubenswrapper[4767]: E0317 16:06:52.816621 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb28438-621e-41a8-b1ad-a07bf5e181c5" containerName="nova-metadata-metadata" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.816742 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb28438-621e-41a8-b1ad-a07bf5e181c5" containerName="nova-metadata-metadata" Mar 17 16:06:52 crc kubenswrapper[4767]: E0317 16:06:52.816829 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb28438-621e-41a8-b1ad-a07bf5e181c5" containerName="nova-metadata-log" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.816899 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb28438-621e-41a8-b1ad-a07bf5e181c5" containerName="nova-metadata-log" Mar 17 16:06:52 crc kubenswrapper[4767]: E0317 16:06:52.818497 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c09859-d250-40a0-97a1-aa8fd1371432" containerName="nova-cell1-novncproxy-novncproxy" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.819820 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c09859-d250-40a0-97a1-aa8fd1371432" containerName="nova-cell1-novncproxy-novncproxy" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.820233 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fb28438-621e-41a8-b1ad-a07bf5e181c5" containerName="nova-metadata-log" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.820345 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="64c09859-d250-40a0-97a1-aa8fd1371432" containerName="nova-cell1-novncproxy-novncproxy" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.820415 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fb28438-621e-41a8-b1ad-a07bf5e181c5" containerName="nova-metadata-metadata" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.823021 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.829183 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.829566 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.829756 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.834676 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.835710 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.835731 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q2mk\" (UniqueName: \"kubernetes.io/projected/b5e808b6-1ae8-4bae-a682-a618234210b1-kube-api-access-5q2mk\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.836017 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.836127 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.836615 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.836852 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjwk2\" (UniqueName: \"kubernetes.io/projected/0fb28438-621e-41a8-b1ad-a07bf5e181c5-kube-api-access-cjwk2\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.836902 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb28438-621e-41a8-b1ad-a07bf5e181c5-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.853633 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.956136 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.956321 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.956425 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q2mk\" (UniqueName: \"kubernetes.io/projected/b5e808b6-1ae8-4bae-a682-a618234210b1-kube-api-access-5q2mk\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.956525 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.956578 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.964354 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.966159 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.967150 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.967903 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5e808b6-1ae8-4bae-a682-a618234210b1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:52 crc kubenswrapper[4767]: I0317 16:06:52.979215 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q2mk\" (UniqueName: \"kubernetes.io/projected/b5e808b6-1ae8-4bae-a682-a618234210b1-kube-api-access-5q2mk\") pod \"nova-cell1-novncproxy-0\" (UID: \"b5e808b6-1ae8-4bae-a682-a618234210b1\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.166969 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.260593 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a1902cc-bb06-4d4e-a90d-b2aa722c3278","Type":"ContainerStarted","Data":"4e71eeaca5a791d0f48e0e70279204c50bd052d87bfc361ebde5e441793f58a9"} Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.267714 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fb28438-621e-41a8-b1ad-a07bf5e181c5","Type":"ContainerDied","Data":"13c22b9e60f48132740d5d63e72b90ad413c9a23e241b579b5a241a12ec7bfd4"} Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.267798 4767 scope.go:117] "RemoveContainer" containerID="361ad6c49e7669c253f3bc43711cc6f6a6b0ce56e8ef019a14e944705c3324eb" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.268051 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.347013 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.399812 4767 scope.go:117] "RemoveContainer" containerID="9d8a897af0b9f6ed3b8a01486cdb6f477b0d3cb7be06b757dc35d75f02e26412" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.432693 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64c09859-d250-40a0-97a1-aa8fd1371432" path="/var/lib/kubelet/pods/64c09859-d250-40a0-97a1-aa8fd1371432/volumes" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.433621 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2925ba9-6448-48b8-8ec0-3fb1830cfc01" path="/var/lib/kubelet/pods/d2925ba9-6448-48b8-8ec0-3fb1830cfc01/volumes" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.449133 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.449216 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.451788 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.455534 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.462845 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.465583 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.478317 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-config-data\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.478460 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.478570 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.478676 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv5td\" (UniqueName: \"kubernetes.io/projected/8a81e7bf-b58a-4421-a981-4a69221f9a5f-kube-api-access-jv5td\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.478705 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a81e7bf-b58a-4421-a981-4a69221f9a5f-logs\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.582303 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-config-data\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.582403 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.582503 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.582593 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv5td\" (UniqueName: \"kubernetes.io/projected/8a81e7bf-b58a-4421-a981-4a69221f9a5f-kube-api-access-jv5td\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.582620 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a81e7bf-b58a-4421-a981-4a69221f9a5f-logs\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.583010 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a81e7bf-b58a-4421-a981-4a69221f9a5f-logs\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.586477 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-config-data\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.588709 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.589774 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.608733 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv5td\" (UniqueName: \"kubernetes.io/projected/8a81e7bf-b58a-4421-a981-4a69221f9a5f-kube-api-access-jv5td\") pod \"nova-metadata-0\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " pod="openstack/nova-metadata-0" Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.798709 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 16:06:53 crc kubenswrapper[4767]: I0317 16:06:53.803677 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 16:06:54 crc kubenswrapper[4767]: I0317 16:06:54.286124 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a1902cc-bb06-4d4e-a90d-b2aa722c3278","Type":"ContainerStarted","Data":"9d09f4f9629834efaee61343aede3c127a42b4eb05c028ac6f2ed85d91d3fb54"} Mar 17 16:06:54 crc kubenswrapper[4767]: I0317 16:06:54.290493 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b5e808b6-1ae8-4bae-a682-a618234210b1","Type":"ContainerStarted","Data":"223ecfcba05473546e8a66f2aed512034f3aa53a639f35d78b7c515b20aa206e"} Mar 17 16:06:54 crc kubenswrapper[4767]: I0317 16:06:54.355885 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:06:54 crc kubenswrapper[4767]: E0317 16:06:54.356266 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:06:54 crc kubenswrapper[4767]: I0317 16:06:54.981820 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:06:55 crc kubenswrapper[4767]: I0317 16:06:55.341004 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a81e7bf-b58a-4421-a981-4a69221f9a5f","Type":"ContainerStarted","Data":"f3cf633b26e09018051f5fe48015ec062e4e687ac9f31c953ac6def5c9069f21"} Mar 17 16:06:55 crc kubenswrapper[4767]: I0317 16:06:55.341536 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a81e7bf-b58a-4421-a981-4a69221f9a5f","Type":"ContainerStarted","Data":"636b97d2519b84768102c0a14a4f7d5b7415c99625dfef25fc09ef40eea89c9f"} Mar 17 16:06:55 crc kubenswrapper[4767]: I0317 16:06:55.345738 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b5e808b6-1ae8-4bae-a682-a618234210b1","Type":"ContainerStarted","Data":"1a9cab848c86c136d097c70bae24278e0f26499b68ee67e66d46f4c6dad150c6"} Mar 17 16:06:55 crc kubenswrapper[4767]: I0317 16:06:55.415640 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.41561264 podStartE2EDuration="3.41561264s" podCreationTimestamp="2026-03-17 16:06:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:06:55.380088363 +0000 UTC m=+1806.793404420" watchObservedRunningTime="2026-03-17 16:06:55.41561264 +0000 UTC m=+1806.828928687" Mar 17 16:06:55 crc kubenswrapper[4767]: I0317 16:06:55.429980 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fb28438-621e-41a8-b1ad-a07bf5e181c5" path="/var/lib/kubelet/pods/0fb28438-621e-41a8-b1ad-a07bf5e181c5/volumes" Mar 17 16:06:56 crc kubenswrapper[4767]: I0317 16:06:56.365632 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a81e7bf-b58a-4421-a981-4a69221f9a5f","Type":"ContainerStarted","Data":"6e117d82f17da24ab40d63740750d5351b29e66934dbcc8001415c37592500dd"} Mar 17 16:06:56 crc kubenswrapper[4767]: I0317 16:06:56.372222 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a1902cc-bb06-4d4e-a90d-b2aa722c3278","Type":"ContainerStarted","Data":"5bfd62d2a37b427f1327e54d79cfbc61be761fe3e0636295d8cc692c895c9726"} Mar 17 16:06:56 crc kubenswrapper[4767]: I0317 16:06:56.372544 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a1902cc-bb06-4d4e-a90d-b2aa722c3278","Type":"ContainerStarted","Data":"d4d23268c4bc05a1440f7f95c40c60364f5d3afdbf9b61f30e15dce5482f3311"} Mar 17 16:06:56 crc kubenswrapper[4767]: I0317 16:06:56.401307 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.401267011 podStartE2EDuration="3.401267011s" podCreationTimestamp="2026-03-17 16:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:06:56.388423884 +0000 UTC m=+1807.801739951" watchObservedRunningTime="2026-03-17 16:06:56.401267011 +0000 UTC m=+1807.814583058" Mar 17 16:06:58 crc kubenswrapper[4767]: I0317 16:06:58.168927 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:06:58 crc kubenswrapper[4767]: I0317 16:06:58.637054 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 17 16:06:58 crc kubenswrapper[4767]: I0317 16:06:58.637705 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 17 16:06:59 crc kubenswrapper[4767]: I0317 16:06:59.416662 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a1902cc-bb06-4d4e-a90d-b2aa722c3278","Type":"ContainerStarted","Data":"c1741b3bbd6661ab98674ec347639322ac63c3e8849da8a860198dc9706f7077"} Mar 17 16:06:59 crc kubenswrapper[4767]: I0317 16:06:59.417051 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 16:06:59 crc kubenswrapper[4767]: I0317 16:06:59.469479 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.996283456 podStartE2EDuration="8.469448666s" podCreationTimestamp="2026-03-17 16:06:51 +0000 UTC" firstStartedPulling="2026-03-17 16:06:52.743827473 +0000 UTC m=+1804.157143520" lastFinishedPulling="2026-03-17 16:06:58.216992683 +0000 UTC m=+1809.630308730" observedRunningTime="2026-03-17 16:06:59.467329885 +0000 UTC m=+1810.880645952" watchObservedRunningTime="2026-03-17 16:06:59.469448666 +0000 UTC m=+1810.882764713" Mar 17 16:07:00 crc kubenswrapper[4767]: I0317 16:07:00.641566 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 17 16:07:00 crc kubenswrapper[4767]: I0317 16:07:00.645873 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 17 16:07:00 crc kubenswrapper[4767]: I0317 16:07:00.646683 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.448957 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.679786 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt"] Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.691770 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.715594 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt"] Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.877235 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.877347 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.877428 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.877581 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.877695 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-config\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.877768 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64z9w\" (UniqueName: \"kubernetes.io/projected/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-kube-api-access-64z9w\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.980058 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.980224 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.980301 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-config\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.980352 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64z9w\" (UniqueName: \"kubernetes.io/projected/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-kube-api-access-64z9w\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.980411 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.980444 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.981086 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.981166 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.981261 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.981429 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-config\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:01 crc kubenswrapper[4767]: I0317 16:07:01.981615 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:02 crc kubenswrapper[4767]: I0317 16:07:02.022392 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64z9w\" (UniqueName: \"kubernetes.io/projected/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-kube-api-access-64z9w\") pod \"dnsmasq-dns-6b7bbf7cf9-tw9qt\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:02 crc kubenswrapper[4767]: I0317 16:07:02.033973 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:03 crc kubenswrapper[4767]: I0317 16:07:03.038386 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt"] Mar 17 16:07:03 crc kubenswrapper[4767]: I0317 16:07:03.169022 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:07:03 crc kubenswrapper[4767]: I0317 16:07:03.220053 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:07:03 crc kubenswrapper[4767]: I0317 16:07:03.493217 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" event={"ID":"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52","Type":"ContainerStarted","Data":"72b84b762715964cba4e0b080485b73617ff7e0a0010833944f58f527307b8f1"} Mar 17 16:07:03 crc kubenswrapper[4767]: I0317 16:07:03.524911 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 17 16:07:03 crc kubenswrapper[4767]: I0317 16:07:03.805619 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 17 16:07:03 crc kubenswrapper[4767]: I0317 16:07:03.805984 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.078268 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-vcxwn"] Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.080558 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.105949 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.106190 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.139547 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-vcxwn"] Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.150274 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-scripts\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.150369 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtk2d\" (UniqueName: \"kubernetes.io/projected/3e2e4698-f52a-476f-8d61-da9459bb6fe0-kube-api-access-mtk2d\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.150472 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-config-data\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.150517 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.253047 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.253508 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-scripts\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.253605 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtk2d\" (UniqueName: \"kubernetes.io/projected/3e2e4698-f52a-476f-8d61-da9459bb6fe0-kube-api-access-mtk2d\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.253730 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-config-data\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.261629 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.261963 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-config-data\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.264147 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-scripts\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.285212 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtk2d\" (UniqueName: \"kubernetes.io/projected/3e2e4698-f52a-476f-8d61-da9459bb6fe0-kube-api-access-mtk2d\") pod \"nova-cell1-cell-mapping-vcxwn\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.436391 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.548012 4767 generic.go:334] "Generic (PLEG): container finished" podID="a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" containerID="0769b432f754f478e1498527b6409bf1a54a6686b7d987b80f89e082fecce414" exitCode=0 Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.549968 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" event={"ID":"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52","Type":"ContainerDied","Data":"0769b432f754f478e1498527b6409bf1a54a6686b7d987b80f89e082fecce414"} Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.844811 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.8:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 16:07:04 crc kubenswrapper[4767]: I0317 16:07:04.845829 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.8:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 16:07:05 crc kubenswrapper[4767]: I0317 16:07:05.215403 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-vcxwn"] Mar 17 16:07:05 crc kubenswrapper[4767]: I0317 16:07:05.690366 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vcxwn" event={"ID":"3e2e4698-f52a-476f-8d61-da9459bb6fe0","Type":"ContainerStarted","Data":"d19686bfa69a34e9069372b37d44d317e8baea7388aeea5299a887afe1e6bd7c"} Mar 17 16:07:05 crc kubenswrapper[4767]: I0317 16:07:05.699808 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" event={"ID":"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52","Type":"ContainerStarted","Data":"dd041368d5d907e35d0a9627903cda8654a62f7b46643599a6bcd662cfa25d67"} Mar 17 16:07:05 crc kubenswrapper[4767]: I0317 16:07:05.701791 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:05 crc kubenswrapper[4767]: I0317 16:07:05.756564 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" podStartSLOduration=4.756528956 podStartE2EDuration="4.756528956s" podCreationTimestamp="2026-03-17 16:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:07:05.738611973 +0000 UTC m=+1817.151928030" watchObservedRunningTime="2026-03-17 16:07:05.756528956 +0000 UTC m=+1817.169845003" Mar 17 16:07:06 crc kubenswrapper[4767]: I0317 16:07:06.295604 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:06 crc kubenswrapper[4767]: I0317 16:07:06.296161 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="85921833-d8a5-4fdf-8687-5607b22b263b" containerName="nova-api-log" containerID="cri-o://101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666" gracePeriod=30 Mar 17 16:07:06 crc kubenswrapper[4767]: I0317 16:07:06.296942 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="85921833-d8a5-4fdf-8687-5607b22b263b" containerName="nova-api-api" containerID="cri-o://a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8" gracePeriod=30 Mar 17 16:07:06 crc kubenswrapper[4767]: I0317 16:07:06.366900 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:07:06 crc kubenswrapper[4767]: E0317 16:07:06.367666 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:07:06 crc kubenswrapper[4767]: I0317 16:07:06.714478 4767 generic.go:334] "Generic (PLEG): container finished" podID="85921833-d8a5-4fdf-8687-5607b22b263b" containerID="101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666" exitCode=143 Mar 17 16:07:06 crc kubenswrapper[4767]: I0317 16:07:06.714566 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85921833-d8a5-4fdf-8687-5607b22b263b","Type":"ContainerDied","Data":"101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666"} Mar 17 16:07:06 crc kubenswrapper[4767]: I0317 16:07:06.717895 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vcxwn" event={"ID":"3e2e4698-f52a-476f-8d61-da9459bb6fe0","Type":"ContainerStarted","Data":"0299cbebda8a41b6539886434bc5f0eded000fbd93450631eedae08c5e159aeb"} Mar 17 16:07:06 crc kubenswrapper[4767]: I0317 16:07:06.757631 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-vcxwn" podStartSLOduration=2.757611119 podStartE2EDuration="2.757611119s" podCreationTimestamp="2026-03-17 16:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:07:06.752339378 +0000 UTC m=+1818.165655425" watchObservedRunningTime="2026-03-17 16:07:06.757611119 +0000 UTC m=+1818.170927166" Mar 17 16:07:07 crc kubenswrapper[4767]: I0317 16:07:07.603635 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:07 crc kubenswrapper[4767]: I0317 16:07:07.604377 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="ceilometer-central-agent" containerID="cri-o://9d09f4f9629834efaee61343aede3c127a42b4eb05c028ac6f2ed85d91d3fb54" gracePeriod=30 Mar 17 16:07:07 crc kubenswrapper[4767]: I0317 16:07:07.604430 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="proxy-httpd" containerID="cri-o://c1741b3bbd6661ab98674ec347639322ac63c3e8849da8a860198dc9706f7077" gracePeriod=30 Mar 17 16:07:07 crc kubenswrapper[4767]: I0317 16:07:07.604531 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="sg-core" containerID="cri-o://5bfd62d2a37b427f1327e54d79cfbc61be761fe3e0636295d8cc692c895c9726" gracePeriod=30 Mar 17 16:07:07 crc kubenswrapper[4767]: I0317 16:07:07.604604 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="ceilometer-notification-agent" containerID="cri-o://d4d23268c4bc05a1440f7f95c40c60364f5d3afdbf9b61f30e15dce5482f3311" gracePeriod=30 Mar 17 16:07:08 crc kubenswrapper[4767]: E0317 16:07:08.718690 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a1902cc_bb06_4d4e_a90d_b2aa722c3278.slice/crio-conmon-d4d23268c4bc05a1440f7f95c40c60364f5d3afdbf9b61f30e15dce5482f3311.scope\": RecentStats: unable to find data in memory cache]" Mar 17 16:07:08 crc kubenswrapper[4767]: I0317 16:07:08.785076 4767 generic.go:334] "Generic (PLEG): container finished" podID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerID="c1741b3bbd6661ab98674ec347639322ac63c3e8849da8a860198dc9706f7077" exitCode=0 Mar 17 16:07:08 crc kubenswrapper[4767]: I0317 16:07:08.785120 4767 generic.go:334] "Generic (PLEG): container finished" podID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerID="5bfd62d2a37b427f1327e54d79cfbc61be761fe3e0636295d8cc692c895c9726" exitCode=2 Mar 17 16:07:08 crc kubenswrapper[4767]: I0317 16:07:08.785129 4767 generic.go:334] "Generic (PLEG): container finished" podID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerID="d4d23268c4bc05a1440f7f95c40c60364f5d3afdbf9b61f30e15dce5482f3311" exitCode=0 Mar 17 16:07:08 crc kubenswrapper[4767]: I0317 16:07:08.785139 4767 generic.go:334] "Generic (PLEG): container finished" podID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerID="9d09f4f9629834efaee61343aede3c127a42b4eb05c028ac6f2ed85d91d3fb54" exitCode=0 Mar 17 16:07:08 crc kubenswrapper[4767]: I0317 16:07:08.785192 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a1902cc-bb06-4d4e-a90d-b2aa722c3278","Type":"ContainerDied","Data":"c1741b3bbd6661ab98674ec347639322ac63c3e8849da8a860198dc9706f7077"} Mar 17 16:07:08 crc kubenswrapper[4767]: I0317 16:07:08.785246 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a1902cc-bb06-4d4e-a90d-b2aa722c3278","Type":"ContainerDied","Data":"5bfd62d2a37b427f1327e54d79cfbc61be761fe3e0636295d8cc692c895c9726"} Mar 17 16:07:08 crc kubenswrapper[4767]: I0317 16:07:08.785261 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a1902cc-bb06-4d4e-a90d-b2aa722c3278","Type":"ContainerDied","Data":"d4d23268c4bc05a1440f7f95c40c60364f5d3afdbf9b61f30e15dce5482f3311"} Mar 17 16:07:08 crc kubenswrapper[4767]: I0317 16:07:08.785272 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a1902cc-bb06-4d4e-a90d-b2aa722c3278","Type":"ContainerDied","Data":"9d09f4f9629834efaee61343aede3c127a42b4eb05c028ac6f2ed85d91d3fb54"} Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.541616 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.650506 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-config-data\") pod \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.650623 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-log-httpd\") pod \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.650680 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-combined-ca-bundle\") pod \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.650771 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-scripts\") pod \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.650835 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kr92\" (UniqueName: \"kubernetes.io/projected/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-kube-api-access-9kr92\") pod \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.650974 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-sg-core-conf-yaml\") pod \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.651017 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-run-httpd\") pod \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\" (UID: \"6a1902cc-bb06-4d4e-a90d-b2aa722c3278\") " Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.652385 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6a1902cc-bb06-4d4e-a90d-b2aa722c3278" (UID: "6a1902cc-bb06-4d4e-a90d-b2aa722c3278"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.653074 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6a1902cc-bb06-4d4e-a90d-b2aa722c3278" (UID: "6a1902cc-bb06-4d4e-a90d-b2aa722c3278"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.661325 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-scripts" (OuterVolumeSpecName: "scripts") pod "6a1902cc-bb06-4d4e-a90d-b2aa722c3278" (UID: "6a1902cc-bb06-4d4e-a90d-b2aa722c3278"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.661993 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-kube-api-access-9kr92" (OuterVolumeSpecName: "kube-api-access-9kr92") pod "6a1902cc-bb06-4d4e-a90d-b2aa722c3278" (UID: "6a1902cc-bb06-4d4e-a90d-b2aa722c3278"). InnerVolumeSpecName "kube-api-access-9kr92". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.722588 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6a1902cc-bb06-4d4e-a90d-b2aa722c3278" (UID: "6a1902cc-bb06-4d4e-a90d-b2aa722c3278"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.754865 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.754962 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kr92\" (UniqueName: \"kubernetes.io/projected/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-kube-api-access-9kr92\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.754979 4767 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.754989 4767 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.755000 4767 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.774339 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a1902cc-bb06-4d4e-a90d-b2aa722c3278" (UID: "6a1902cc-bb06-4d4e-a90d-b2aa722c3278"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.801833 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a1902cc-bb06-4d4e-a90d-b2aa722c3278","Type":"ContainerDied","Data":"4e71eeaca5a791d0f48e0e70279204c50bd052d87bfc361ebde5e441793f58a9"} Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.801910 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.801958 4767 scope.go:117] "RemoveContainer" containerID="c1741b3bbd6661ab98674ec347639322ac63c3e8849da8a860198dc9706f7077" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.832507 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-config-data" (OuterVolumeSpecName: "config-data") pod "6a1902cc-bb06-4d4e-a90d-b2aa722c3278" (UID: "6a1902cc-bb06-4d4e-a90d-b2aa722c3278"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.858030 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.858093 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1902cc-bb06-4d4e-a90d-b2aa722c3278-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:09 crc kubenswrapper[4767]: I0317 16:07:09.994230 4767 scope.go:117] "RemoveContainer" containerID="5bfd62d2a37b427f1327e54d79cfbc61be761fe3e0636295d8cc692c895c9726" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.027938 4767 scope.go:117] "RemoveContainer" containerID="d4d23268c4bc05a1440f7f95c40c60364f5d3afdbf9b61f30e15dce5482f3311" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.063903 4767 scope.go:117] "RemoveContainer" containerID="9d09f4f9629834efaee61343aede3c127a42b4eb05c028ac6f2ed85d91d3fb54" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.174261 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.194772 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.211274 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:10 crc kubenswrapper[4767]: E0317 16:07:10.212082 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="ceilometer-notification-agent" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.212110 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="ceilometer-notification-agent" Mar 17 16:07:10 crc kubenswrapper[4767]: E0317 16:07:10.212332 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="sg-core" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.212351 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="sg-core" Mar 17 16:07:10 crc kubenswrapper[4767]: E0317 16:07:10.212364 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="ceilometer-central-agent" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.212375 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="ceilometer-central-agent" Mar 17 16:07:10 crc kubenswrapper[4767]: E0317 16:07:10.212403 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="proxy-httpd" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.212413 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="proxy-httpd" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.212793 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="ceilometer-notification-agent" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.212815 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="sg-core" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.212832 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="ceilometer-central-agent" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.212864 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" containerName="proxy-httpd" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.216042 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.224544 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.228940 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.229488 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.271255 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.277775 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-log-httpd\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.278060 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.278134 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn7kt\" (UniqueName: \"kubernetes.io/projected/c74426cc-3c82-4459-b9ab-ba39730f182a-kube-api-access-wn7kt\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.278556 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.278788 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-run-httpd\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.279062 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-scripts\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.279235 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-config-data\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.380839 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-config-data\") pod \"85921833-d8a5-4fdf-8687-5607b22b263b\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.380994 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-combined-ca-bundle\") pod \"85921833-d8a5-4fdf-8687-5607b22b263b\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.381023 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tc5lq\" (UniqueName: \"kubernetes.io/projected/85921833-d8a5-4fdf-8687-5607b22b263b-kube-api-access-tc5lq\") pod \"85921833-d8a5-4fdf-8687-5607b22b263b\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.381250 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85921833-d8a5-4fdf-8687-5607b22b263b-logs\") pod \"85921833-d8a5-4fdf-8687-5607b22b263b\" (UID: \"85921833-d8a5-4fdf-8687-5607b22b263b\") " Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.381621 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-log-httpd\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.381723 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.381754 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn7kt\" (UniqueName: \"kubernetes.io/projected/c74426cc-3c82-4459-b9ab-ba39730f182a-kube-api-access-wn7kt\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.381810 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.381866 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-run-httpd\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.381943 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-scripts\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.381999 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-config-data\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.383879 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-log-httpd\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.391108 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-run-httpd\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.391262 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-config-data\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.394035 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85921833-d8a5-4fdf-8687-5607b22b263b-logs" (OuterVolumeSpecName: "logs") pod "85921833-d8a5-4fdf-8687-5607b22b263b" (UID: "85921833-d8a5-4fdf-8687-5607b22b263b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.397148 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85921833-d8a5-4fdf-8687-5607b22b263b-kube-api-access-tc5lq" (OuterVolumeSpecName: "kube-api-access-tc5lq") pod "85921833-d8a5-4fdf-8687-5607b22b263b" (UID: "85921833-d8a5-4fdf-8687-5607b22b263b"). InnerVolumeSpecName "kube-api-access-tc5lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.404149 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.406438 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-scripts\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.419993 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.431159 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn7kt\" (UniqueName: \"kubernetes.io/projected/c74426cc-3c82-4459-b9ab-ba39730f182a-kube-api-access-wn7kt\") pod \"ceilometer-0\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.444415 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85921833-d8a5-4fdf-8687-5607b22b263b" (UID: "85921833-d8a5-4fdf-8687-5607b22b263b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.486102 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85921833-d8a5-4fdf-8687-5607b22b263b-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.486141 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tc5lq\" (UniqueName: \"kubernetes.io/projected/85921833-d8a5-4fdf-8687-5607b22b263b-kube-api-access-tc5lq\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.486158 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.487882 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-config-data" (OuterVolumeSpecName: "config-data") pod "85921833-d8a5-4fdf-8687-5607b22b263b" (UID: "85921833-d8a5-4fdf-8687-5607b22b263b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.554671 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.588607 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85921833-d8a5-4fdf-8687-5607b22b263b-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.847662 4767 generic.go:334] "Generic (PLEG): container finished" podID="85921833-d8a5-4fdf-8687-5607b22b263b" containerID="a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8" exitCode=0 Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.848036 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85921833-d8a5-4fdf-8687-5607b22b263b","Type":"ContainerDied","Data":"a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8"} Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.848074 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85921833-d8a5-4fdf-8687-5607b22b263b","Type":"ContainerDied","Data":"f67473e2e8fc943cc67c322282a752054e4318da10b1d0037a307255d1f4e03a"} Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.848099 4767 scope.go:117] "RemoveContainer" containerID="a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.848311 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.893552 4767 scope.go:117] "RemoveContainer" containerID="101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.930614 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.952620 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.964588 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:10 crc kubenswrapper[4767]: E0317 16:07:10.965201 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85921833-d8a5-4fdf-8687-5607b22b263b" containerName="nova-api-log" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.965224 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="85921833-d8a5-4fdf-8687-5607b22b263b" containerName="nova-api-log" Mar 17 16:07:10 crc kubenswrapper[4767]: E0317 16:07:10.965261 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85921833-d8a5-4fdf-8687-5607b22b263b" containerName="nova-api-api" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.965270 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="85921833-d8a5-4fdf-8687-5607b22b263b" containerName="nova-api-api" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.965548 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="85921833-d8a5-4fdf-8687-5607b22b263b" containerName="nova-api-log" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.965565 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="85921833-d8a5-4fdf-8687-5607b22b263b" containerName="nova-api-api" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.967008 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.969391 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.969810 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.969951 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.987667 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.995009 4767 scope.go:117] "RemoveContainer" containerID="a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8" Mar 17 16:07:10 crc kubenswrapper[4767]: E0317 16:07:10.995812 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8\": container with ID starting with a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8 not found: ID does not exist" containerID="a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.995866 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8"} err="failed to get container status \"a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8\": rpc error: code = NotFound desc = could not find container \"a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8\": container with ID starting with a5ffc859eabedbf1ff49f6d745fed519c0abd1e2cdb612326c2a6c55d6f5def8 not found: ID does not exist" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.995897 4767 scope.go:117] "RemoveContainer" containerID="101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666" Mar 17 16:07:10 crc kubenswrapper[4767]: E0317 16:07:10.997057 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666\": container with ID starting with 101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666 not found: ID does not exist" containerID="101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666" Mar 17 16:07:10 crc kubenswrapper[4767]: I0317 16:07:10.997117 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666"} err="failed to get container status \"101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666\": rpc error: code = NotFound desc = could not find container \"101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666\": container with ID starting with 101d872196adda0ac4b2c59a22e7155a26257acaffa058b9efd4d93a2299c666 not found: ID does not exist" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.105688 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.105791 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.105830 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-config-data\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.105896 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7br2j\" (UniqueName: \"kubernetes.io/projected/a108be47-f28c-4b0e-985b-04ffdbecf880-kube-api-access-7br2j\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.105983 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-public-tls-certs\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.106233 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a108be47-f28c-4b0e-985b-04ffdbecf880-logs\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.130785 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.190541 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.209062 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.209300 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.209332 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-config-data\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.209381 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7br2j\" (UniqueName: \"kubernetes.io/projected/a108be47-f28c-4b0e-985b-04ffdbecf880-kube-api-access-7br2j\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.209426 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-public-tls-certs\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.209536 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a108be47-f28c-4b0e-985b-04ffdbecf880-logs\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.210078 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a108be47-f28c-4b0e-985b-04ffdbecf880-logs\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.216656 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.217534 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.217539 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-public-tls-certs\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.217888 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-config-data\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.234567 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7br2j\" (UniqueName: \"kubernetes.io/projected/a108be47-f28c-4b0e-985b-04ffdbecf880-kube-api-access-7br2j\") pod \"nova-api-0\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.306791 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.384512 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a1902cc-bb06-4d4e-a90d-b2aa722c3278" path="/var/lib/kubelet/pods/6a1902cc-bb06-4d4e-a90d-b2aa722c3278/volumes" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.385777 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85921833-d8a5-4fdf-8687-5607b22b263b" path="/var/lib/kubelet/pods/85921833-d8a5-4fdf-8687-5607b22b263b/volumes" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.805556 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.808329 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 17 16:07:11 crc kubenswrapper[4767]: I0317 16:07:11.875525 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74426cc-3c82-4459-b9ab-ba39730f182a","Type":"ContainerStarted","Data":"c3bc1821db2037b1eb67e35b95149b6081e5d93429037354563100af08e50c49"} Mar 17 16:07:12 crc kubenswrapper[4767]: I0317 16:07:12.256603 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:07:12 crc kubenswrapper[4767]: I0317 16:07:12.336254 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:12 crc kubenswrapper[4767]: I0317 16:07:12.355387 4767 scope.go:117] "RemoveContainer" containerID="cc06b7dcb8534812e686a7f25518ad0d196b5c0386841ca0064c5eaba1dc1848" Mar 17 16:07:12 crc kubenswrapper[4767]: I0317 16:07:12.424302 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-z9szv"] Mar 17 16:07:12 crc kubenswrapper[4767]: I0317 16:07:12.424660 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" podUID="f3467dd4-3ba8-4492-a678-611b6b982946" containerName="dnsmasq-dns" containerID="cri-o://f8b5447b2f04eaa526827541b47e97798418eb8585f7249cf7d8889d5722b52b" gracePeriod=10 Mar 17 16:07:12 crc kubenswrapper[4767]: I0317 16:07:12.933586 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74426cc-3c82-4459-b9ab-ba39730f182a","Type":"ContainerStarted","Data":"badd6c9149b8055fe6ed0453e3037e93998c97262d51784b281eb766d15262a9"} Mar 17 16:07:12 crc kubenswrapper[4767]: I0317 16:07:12.944410 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a108be47-f28c-4b0e-985b-04ffdbecf880","Type":"ContainerStarted","Data":"0691bbc5bde42dfdec202493086bba02f881dbfe5ca0aee08ba6d976da9dfedb"} Mar 17 16:07:12 crc kubenswrapper[4767]: I0317 16:07:12.944981 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a108be47-f28c-4b0e-985b-04ffdbecf880","Type":"ContainerStarted","Data":"f9ed557d8b2effaacc15062cb612019a8e4ba586ff1116bb09fb03c3949b84ae"} Mar 17 16:07:12 crc kubenswrapper[4767]: I0317 16:07:12.960359 4767 generic.go:334] "Generic (PLEG): container finished" podID="f3467dd4-3ba8-4492-a678-611b6b982946" containerID="f8b5447b2f04eaa526827541b47e97798418eb8585f7249cf7d8889d5722b52b" exitCode=0 Mar 17 16:07:12 crc kubenswrapper[4767]: I0317 16:07:12.960422 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" event={"ID":"f3467dd4-3ba8-4492-a678-611b6b982946","Type":"ContainerDied","Data":"f8b5447b2f04eaa526827541b47e97798418eb8585f7249cf7d8889d5722b52b"} Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.120051 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.303140 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-svc\") pod \"f3467dd4-3ba8-4492-a678-611b6b982946\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.304525 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-config\") pod \"f3467dd4-3ba8-4492-a678-611b6b982946\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.304581 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-sb\") pod \"f3467dd4-3ba8-4492-a678-611b6b982946\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.304659 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-swift-storage-0\") pod \"f3467dd4-3ba8-4492-a678-611b6b982946\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.304863 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-nb\") pod \"f3467dd4-3ba8-4492-a678-611b6b982946\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.304933 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrvp8\" (UniqueName: \"kubernetes.io/projected/f3467dd4-3ba8-4492-a678-611b6b982946-kube-api-access-zrvp8\") pod \"f3467dd4-3ba8-4492-a678-611b6b982946\" (UID: \"f3467dd4-3ba8-4492-a678-611b6b982946\") " Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.321826 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3467dd4-3ba8-4492-a678-611b6b982946-kube-api-access-zrvp8" (OuterVolumeSpecName: "kube-api-access-zrvp8") pod "f3467dd4-3ba8-4492-a678-611b6b982946" (UID: "f3467dd4-3ba8-4492-a678-611b6b982946"). InnerVolumeSpecName "kube-api-access-zrvp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.410786 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrvp8\" (UniqueName: \"kubernetes.io/projected/f3467dd4-3ba8-4492-a678-611b6b982946-kube-api-access-zrvp8\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.525237 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f3467dd4-3ba8-4492-a678-611b6b982946" (UID: "f3467dd4-3ba8-4492-a678-611b6b982946"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.537942 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f3467dd4-3ba8-4492-a678-611b6b982946" (UID: "f3467dd4-3ba8-4492-a678-611b6b982946"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.549481 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f3467dd4-3ba8-4492-a678-611b6b982946" (UID: "f3467dd4-3ba8-4492-a678-611b6b982946"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.551115 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-config" (OuterVolumeSpecName: "config") pod "f3467dd4-3ba8-4492-a678-611b6b982946" (UID: "f3467dd4-3ba8-4492-a678-611b6b982946"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.562059 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f3467dd4-3ba8-4492-a678-611b6b982946" (UID: "f3467dd4-3ba8-4492-a678-611b6b982946"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.627650 4767 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.627705 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.627721 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.627734 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.627751 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3467dd4-3ba8-4492-a678-611b6b982946-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.821568 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.837735 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 17 16:07:13 crc kubenswrapper[4767]: I0317 16:07:13.843612 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 17 16:07:14 crc kubenswrapper[4767]: I0317 16:07:14.006867 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74426cc-3c82-4459-b9ab-ba39730f182a","Type":"ContainerStarted","Data":"cc6095ba10d1bedc84ab4b3e688571375cf038de71aac3f583b5df9dda17fcbc"} Mar 17 16:07:14 crc kubenswrapper[4767]: I0317 16:07:14.013914 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a108be47-f28c-4b0e-985b-04ffdbecf880","Type":"ContainerStarted","Data":"30a882e8ff6de652846266fdbaa65c3ce1ede5bb1a5f6e4c2de39f25f6654fa9"} Mar 17 16:07:14 crc kubenswrapper[4767]: I0317 16:07:14.019693 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" Mar 17 16:07:14 crc kubenswrapper[4767]: I0317 16:07:14.019688 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-z9szv" event={"ID":"f3467dd4-3ba8-4492-a678-611b6b982946","Type":"ContainerDied","Data":"47263d72b689335533cdad3559966dedcda33ce22076eb1ddf963a2c426433b7"} Mar 17 16:07:14 crc kubenswrapper[4767]: I0317 16:07:14.019895 4767 scope.go:117] "RemoveContainer" containerID="f8b5447b2f04eaa526827541b47e97798418eb8585f7249cf7d8889d5722b52b" Mar 17 16:07:14 crc kubenswrapper[4767]: I0317 16:07:14.037673 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 17 16:07:14 crc kubenswrapper[4767]: I0317 16:07:14.067672 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.067645707 podStartE2EDuration="4.067645707s" podCreationTimestamp="2026-03-17 16:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:07:14.054623265 +0000 UTC m=+1825.467939332" watchObservedRunningTime="2026-03-17 16:07:14.067645707 +0000 UTC m=+1825.480961754" Mar 17 16:07:14 crc kubenswrapper[4767]: I0317 16:07:14.134227 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-z9szv"] Mar 17 16:07:14 crc kubenswrapper[4767]: I0317 16:07:14.161966 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-z9szv"] Mar 17 16:07:14 crc kubenswrapper[4767]: I0317 16:07:14.212566 4767 scope.go:117] "RemoveContainer" containerID="d33d736e32fc1fe05a1feabe53c8ce8244cb942ee434c5d90cfc10bf73e9ba89" Mar 17 16:07:15 crc kubenswrapper[4767]: I0317 16:07:15.038833 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74426cc-3c82-4459-b9ab-ba39730f182a","Type":"ContainerStarted","Data":"b3e6754c1194ec22767e5bdf230d7f6e7b1e92fdc39e15b6f061f7ad326e0701"} Mar 17 16:07:15 crc kubenswrapper[4767]: I0317 16:07:15.583646 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3467dd4-3ba8-4492-a678-611b6b982946" path="/var/lib/kubelet/pods/f3467dd4-3ba8-4492-a678-611b6b982946/volumes" Mar 17 16:07:16 crc kubenswrapper[4767]: I0317 16:07:16.056437 4767 generic.go:334] "Generic (PLEG): container finished" podID="3e2e4698-f52a-476f-8d61-da9459bb6fe0" containerID="0299cbebda8a41b6539886434bc5f0eded000fbd93450631eedae08c5e159aeb" exitCode=0 Mar 17 16:07:16 crc kubenswrapper[4767]: I0317 16:07:16.056564 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vcxwn" event={"ID":"3e2e4698-f52a-476f-8d61-da9459bb6fe0","Type":"ContainerDied","Data":"0299cbebda8a41b6539886434bc5f0eded000fbd93450631eedae08c5e159aeb"} Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.391773 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:07:17 crc kubenswrapper[4767]: E0317 16:07:17.392387 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.756769 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.840241 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-scripts\") pod \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.840490 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-config-data\") pod \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.848486 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-scripts" (OuterVolumeSpecName: "scripts") pod "3e2e4698-f52a-476f-8d61-da9459bb6fe0" (UID: "3e2e4698-f52a-476f-8d61-da9459bb6fe0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.883786 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-config-data" (OuterVolumeSpecName: "config-data") pod "3e2e4698-f52a-476f-8d61-da9459bb6fe0" (UID: "3e2e4698-f52a-476f-8d61-da9459bb6fe0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.943326 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtk2d\" (UniqueName: \"kubernetes.io/projected/3e2e4698-f52a-476f-8d61-da9459bb6fe0-kube-api-access-mtk2d\") pod \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.943479 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-combined-ca-bundle\") pod \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\" (UID: \"3e2e4698-f52a-476f-8d61-da9459bb6fe0\") " Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.944249 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.944271 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.946887 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e2e4698-f52a-476f-8d61-da9459bb6fe0-kube-api-access-mtk2d" (OuterVolumeSpecName: "kube-api-access-mtk2d") pod "3e2e4698-f52a-476f-8d61-da9459bb6fe0" (UID: "3e2e4698-f52a-476f-8d61-da9459bb6fe0"). InnerVolumeSpecName "kube-api-access-mtk2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:07:17 crc kubenswrapper[4767]: I0317 16:07:17.986512 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e2e4698-f52a-476f-8d61-da9459bb6fe0" (UID: "3e2e4698-f52a-476f-8d61-da9459bb6fe0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.046317 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtk2d\" (UniqueName: \"kubernetes.io/projected/3e2e4698-f52a-476f-8d61-da9459bb6fe0-kube-api-access-mtk2d\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.046704 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2e4698-f52a-476f-8d61-da9459bb6fe0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.084239 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vcxwn" event={"ID":"3e2e4698-f52a-476f-8d61-da9459bb6fe0","Type":"ContainerDied","Data":"d19686bfa69a34e9069372b37d44d317e8baea7388aeea5299a887afe1e6bd7c"} Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.084307 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d19686bfa69a34e9069372b37d44d317e8baea7388aeea5299a887afe1e6bd7c" Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.084268 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vcxwn" Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.087427 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74426cc-3c82-4459-b9ab-ba39730f182a","Type":"ContainerStarted","Data":"46765c1bf5a2acbafd3d6c783f2fdb5d3eaceea216a84bad457961ede06ce025"} Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.087670 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="ceilometer-central-agent" containerID="cri-o://badd6c9149b8055fe6ed0453e3037e93998c97262d51784b281eb766d15262a9" gracePeriod=30 Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.088054 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.088819 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="proxy-httpd" containerID="cri-o://46765c1bf5a2acbafd3d6c783f2fdb5d3eaceea216a84bad457961ede06ce025" gracePeriod=30 Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.088891 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="sg-core" containerID="cri-o://b3e6754c1194ec22767e5bdf230d7f6e7b1e92fdc39e15b6f061f7ad326e0701" gracePeriod=30 Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.088964 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="ceilometer-notification-agent" containerID="cri-o://cc6095ba10d1bedc84ab4b3e688571375cf038de71aac3f583b5df9dda17fcbc" gracePeriod=30 Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.121269 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.810840158 podStartE2EDuration="8.121241996s" podCreationTimestamp="2026-03-17 16:07:10 +0000 UTC" firstStartedPulling="2026-03-17 16:07:11.191658029 +0000 UTC m=+1822.604974076" lastFinishedPulling="2026-03-17 16:07:17.502059867 +0000 UTC m=+1828.915375914" observedRunningTime="2026-03-17 16:07:18.109811459 +0000 UTC m=+1829.523127516" watchObservedRunningTime="2026-03-17 16:07:18.121241996 +0000 UTC m=+1829.534558033" Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.300738 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.301163 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a108be47-f28c-4b0e-985b-04ffdbecf880" containerName="nova-api-log" containerID="cri-o://0691bbc5bde42dfdec202493086bba02f881dbfe5ca0aee08ba6d976da9dfedb" gracePeriod=30 Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.301284 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a108be47-f28c-4b0e-985b-04ffdbecf880" containerName="nova-api-api" containerID="cri-o://30a882e8ff6de652846266fdbaa65c3ce1ede5bb1a5f6e4c2de39f25f6654fa9" gracePeriod=30 Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.319238 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.319653 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2e690f7d-0a21-4296-b677-3c30f7cb878a" containerName="nova-scheduler-scheduler" containerID="cri-o://dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926" gracePeriod=30 Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.337964 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.338275 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerName="nova-metadata-log" containerID="cri-o://f3cf633b26e09018051f5fe48015ec062e4e687ac9f31c953ac6def5c9069f21" gracePeriod=30 Mar 17 16:07:18 crc kubenswrapper[4767]: I0317 16:07:18.338873 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerName="nova-metadata-metadata" containerID="cri-o://6e117d82f17da24ab40d63740750d5351b29e66934dbcc8001415c37592500dd" gracePeriod=30 Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.105615 4767 generic.go:334] "Generic (PLEG): container finished" podID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerID="46765c1bf5a2acbafd3d6c783f2fdb5d3eaceea216a84bad457961ede06ce025" exitCode=0 Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.105910 4767 generic.go:334] "Generic (PLEG): container finished" podID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerID="b3e6754c1194ec22767e5bdf230d7f6e7b1e92fdc39e15b6f061f7ad326e0701" exitCode=2 Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.105919 4767 generic.go:334] "Generic (PLEG): container finished" podID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerID="cc6095ba10d1bedc84ab4b3e688571375cf038de71aac3f583b5df9dda17fcbc" exitCode=0 Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.105684 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74426cc-3c82-4459-b9ab-ba39730f182a","Type":"ContainerDied","Data":"46765c1bf5a2acbafd3d6c783f2fdb5d3eaceea216a84bad457961ede06ce025"} Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.105994 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74426cc-3c82-4459-b9ab-ba39730f182a","Type":"ContainerDied","Data":"b3e6754c1194ec22767e5bdf230d7f6e7b1e92fdc39e15b6f061f7ad326e0701"} Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.106009 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74426cc-3c82-4459-b9ab-ba39730f182a","Type":"ContainerDied","Data":"cc6095ba10d1bedc84ab4b3e688571375cf038de71aac3f583b5df9dda17fcbc"} Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.108967 4767 generic.go:334] "Generic (PLEG): container finished" podID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerID="f3cf633b26e09018051f5fe48015ec062e4e687ac9f31c953ac6def5c9069f21" exitCode=143 Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.109022 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a81e7bf-b58a-4421-a981-4a69221f9a5f","Type":"ContainerDied","Data":"f3cf633b26e09018051f5fe48015ec062e4e687ac9f31c953ac6def5c9069f21"} Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.111971 4767 generic.go:334] "Generic (PLEG): container finished" podID="a108be47-f28c-4b0e-985b-04ffdbecf880" containerID="30a882e8ff6de652846266fdbaa65c3ce1ede5bb1a5f6e4c2de39f25f6654fa9" exitCode=0 Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.112056 4767 generic.go:334] "Generic (PLEG): container finished" podID="a108be47-f28c-4b0e-985b-04ffdbecf880" containerID="0691bbc5bde42dfdec202493086bba02f881dbfe5ca0aee08ba6d976da9dfedb" exitCode=143 Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.112031 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a108be47-f28c-4b0e-985b-04ffdbecf880","Type":"ContainerDied","Data":"30a882e8ff6de652846266fdbaa65c3ce1ede5bb1a5f6e4c2de39f25f6654fa9"} Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.112228 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a108be47-f28c-4b0e-985b-04ffdbecf880","Type":"ContainerDied","Data":"0691bbc5bde42dfdec202493086bba02f881dbfe5ca0aee08ba6d976da9dfedb"} Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.112300 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a108be47-f28c-4b0e-985b-04ffdbecf880","Type":"ContainerDied","Data":"f9ed557d8b2effaacc15062cb612019a8e4ba586ff1116bb09fb03c3949b84ae"} Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.112364 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9ed557d8b2effaacc15062cb612019a8e4ba586ff1116bb09fb03c3949b84ae" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.144731 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.317576 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a108be47-f28c-4b0e-985b-04ffdbecf880-logs\") pod \"a108be47-f28c-4b0e-985b-04ffdbecf880\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.318012 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a108be47-f28c-4b0e-985b-04ffdbecf880-logs" (OuterVolumeSpecName: "logs") pod "a108be47-f28c-4b0e-985b-04ffdbecf880" (UID: "a108be47-f28c-4b0e-985b-04ffdbecf880"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.318118 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-public-tls-certs\") pod \"a108be47-f28c-4b0e-985b-04ffdbecf880\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.318226 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-config-data\") pod \"a108be47-f28c-4b0e-985b-04ffdbecf880\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.318415 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-internal-tls-certs\") pod \"a108be47-f28c-4b0e-985b-04ffdbecf880\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.318954 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-combined-ca-bundle\") pod \"a108be47-f28c-4b0e-985b-04ffdbecf880\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.319279 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7br2j\" (UniqueName: \"kubernetes.io/projected/a108be47-f28c-4b0e-985b-04ffdbecf880-kube-api-access-7br2j\") pod \"a108be47-f28c-4b0e-985b-04ffdbecf880\" (UID: \"a108be47-f28c-4b0e-985b-04ffdbecf880\") " Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.320375 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a108be47-f28c-4b0e-985b-04ffdbecf880-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.325380 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a108be47-f28c-4b0e-985b-04ffdbecf880-kube-api-access-7br2j" (OuterVolumeSpecName: "kube-api-access-7br2j") pod "a108be47-f28c-4b0e-985b-04ffdbecf880" (UID: "a108be47-f28c-4b0e-985b-04ffdbecf880"). InnerVolumeSpecName "kube-api-access-7br2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.352524 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a108be47-f28c-4b0e-985b-04ffdbecf880" (UID: "a108be47-f28c-4b0e-985b-04ffdbecf880"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.367776 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-config-data" (OuterVolumeSpecName: "config-data") pod "a108be47-f28c-4b0e-985b-04ffdbecf880" (UID: "a108be47-f28c-4b0e-985b-04ffdbecf880"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.416674 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a108be47-f28c-4b0e-985b-04ffdbecf880" (UID: "a108be47-f28c-4b0e-985b-04ffdbecf880"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.418811 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a108be47-f28c-4b0e-985b-04ffdbecf880" (UID: "a108be47-f28c-4b0e-985b-04ffdbecf880"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.423543 4767 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.423581 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.423594 4767 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.423610 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a108be47-f28c-4b0e-985b-04ffdbecf880-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:19 crc kubenswrapper[4767]: I0317 16:07:19.423627 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7br2j\" (UniqueName: \"kubernetes.io/projected/a108be47-f28c-4b0e-985b-04ffdbecf880-kube-api-access-7br2j\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.139796 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.189795 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.215335 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.226545 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:20 crc kubenswrapper[4767]: E0317 16:07:20.252637 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3467dd4-3ba8-4492-a678-611b6b982946" containerName="init" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.252673 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3467dd4-3ba8-4492-a678-611b6b982946" containerName="init" Mar 17 16:07:20 crc kubenswrapper[4767]: E0317 16:07:20.252702 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3467dd4-3ba8-4492-a678-611b6b982946" containerName="dnsmasq-dns" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.252709 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3467dd4-3ba8-4492-a678-611b6b982946" containerName="dnsmasq-dns" Mar 17 16:07:20 crc kubenswrapper[4767]: E0317 16:07:20.252729 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e2e4698-f52a-476f-8d61-da9459bb6fe0" containerName="nova-manage" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.252735 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e2e4698-f52a-476f-8d61-da9459bb6fe0" containerName="nova-manage" Mar 17 16:07:20 crc kubenswrapper[4767]: E0317 16:07:20.252777 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a108be47-f28c-4b0e-985b-04ffdbecf880" containerName="nova-api-api" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.252784 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a108be47-f28c-4b0e-985b-04ffdbecf880" containerName="nova-api-api" Mar 17 16:07:20 crc kubenswrapper[4767]: E0317 16:07:20.252799 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a108be47-f28c-4b0e-985b-04ffdbecf880" containerName="nova-api-log" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.252806 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a108be47-f28c-4b0e-985b-04ffdbecf880" containerName="nova-api-log" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.253240 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e2e4698-f52a-476f-8d61-da9459bb6fe0" containerName="nova-manage" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.253253 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3467dd4-3ba8-4492-a678-611b6b982946" containerName="dnsmasq-dns" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.253274 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a108be47-f28c-4b0e-985b-04ffdbecf880" containerName="nova-api-api" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.253295 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a108be47-f28c-4b0e-985b-04ffdbecf880" containerName="nova-api-log" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.254661 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.254755 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.257471 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.258315 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.258336 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.350137 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-config-data\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.350407 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-public-tls-certs\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.350478 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.350590 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.350703 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxqmv\" (UniqueName: \"kubernetes.io/projected/2e3e3d43-9308-4bb7-8046-adf72dda25fe-kube-api-access-qxqmv\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.351088 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e3e3d43-9308-4bb7-8046-adf72dda25fe-logs\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.454553 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e3e3d43-9308-4bb7-8046-adf72dda25fe-logs\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.454809 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-config-data\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.454918 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.454939 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-public-tls-certs\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.455122 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.455229 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxqmv\" (UniqueName: \"kubernetes.io/projected/2e3e3d43-9308-4bb7-8046-adf72dda25fe-kube-api-access-qxqmv\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.459418 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e3e3d43-9308-4bb7-8046-adf72dda25fe-logs\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.465216 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.465812 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-public-tls-certs\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.466516 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-config-data\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.468152 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3e3d43-9308-4bb7-8046-adf72dda25fe-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.479689 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxqmv\" (UniqueName: \"kubernetes.io/projected/2e3e3d43-9308-4bb7-8046-adf72dda25fe-kube-api-access-qxqmv\") pod \"nova-api-0\" (UID: \"2e3e3d43-9308-4bb7-8046-adf72dda25fe\") " pod="openstack/nova-api-0" Mar 17 16:07:20 crc kubenswrapper[4767]: I0317 16:07:20.575853 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.130478 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 16:07:21 crc kubenswrapper[4767]: W0317 16:07:21.158166 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e3e3d43_9308_4bb7_8046_adf72dda25fe.slice/crio-3ae3af5b71e3e924b44bbc4abedac6e5d23c25d2b728d078a99bb4197043856b WatchSource:0}: Error finding container 3ae3af5b71e3e924b44bbc4abedac6e5d23c25d2b728d078a99bb4197043856b: Status 404 returned error can't find the container with id 3ae3af5b71e3e924b44bbc4abedac6e5d23c25d2b728d078a99bb4197043856b Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.175839 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.190488 4767 generic.go:334] "Generic (PLEG): container finished" podID="2e690f7d-0a21-4296-b677-3c30f7cb878a" containerID="dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926" exitCode=0 Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.190563 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.190593 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e690f7d-0a21-4296-b677-3c30f7cb878a","Type":"ContainerDied","Data":"dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926"} Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.190671 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e690f7d-0a21-4296-b677-3c30f7cb878a","Type":"ContainerDied","Data":"b6c47559be33d46a654ed9df5f211bcfb4c6bec895532c836357ab58cfd3ae1e"} Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.190695 4767 scope.go:117] "RemoveContainer" containerID="dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.204053 4767 generic.go:334] "Generic (PLEG): container finished" podID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerID="badd6c9149b8055fe6ed0453e3037e93998c97262d51784b281eb766d15262a9" exitCode=0 Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.204095 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74426cc-3c82-4459-b9ab-ba39730f182a","Type":"ContainerDied","Data":"badd6c9149b8055fe6ed0453e3037e93998c97262d51784b281eb766d15262a9"} Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.252406 4767 scope.go:117] "RemoveContainer" containerID="dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926" Mar 17 16:07:21 crc kubenswrapper[4767]: E0317 16:07:21.252944 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926\": container with ID starting with dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926 not found: ID does not exist" containerID="dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.252985 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926"} err="failed to get container status \"dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926\": rpc error: code = NotFound desc = could not find container \"dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926\": container with ID starting with dfcdbbdbba33139dcd2e233ed1d2cf13d9d7b2ed21c4c673b8b990523ebbb926 not found: ID does not exist" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.294270 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qnxt\" (UniqueName: \"kubernetes.io/projected/2e690f7d-0a21-4296-b677-3c30f7cb878a-kube-api-access-6qnxt\") pod \"2e690f7d-0a21-4296-b677-3c30f7cb878a\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.294716 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-config-data\") pod \"2e690f7d-0a21-4296-b677-3c30f7cb878a\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.295113 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-combined-ca-bundle\") pod \"2e690f7d-0a21-4296-b677-3c30f7cb878a\" (UID: \"2e690f7d-0a21-4296-b677-3c30f7cb878a\") " Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.304262 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e690f7d-0a21-4296-b677-3c30f7cb878a-kube-api-access-6qnxt" (OuterVolumeSpecName: "kube-api-access-6qnxt") pod "2e690f7d-0a21-4296-b677-3c30f7cb878a" (UID: "2e690f7d-0a21-4296-b677-3c30f7cb878a"). InnerVolumeSpecName "kube-api-access-6qnxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.344913 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e690f7d-0a21-4296-b677-3c30f7cb878a" (UID: "2e690f7d-0a21-4296-b677-3c30f7cb878a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.358029 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-config-data" (OuterVolumeSpecName: "config-data") pod "2e690f7d-0a21-4296-b677-3c30f7cb878a" (UID: "2e690f7d-0a21-4296-b677-3c30f7cb878a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.378988 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a108be47-f28c-4b0e-985b-04ffdbecf880" path="/var/lib/kubelet/pods/a108be47-f28c-4b0e-985b-04ffdbecf880/volumes" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.398431 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.398471 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qnxt\" (UniqueName: \"kubernetes.io/projected/2e690f7d-0a21-4296-b677-3c30f7cb878a-kube-api-access-6qnxt\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.398500 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e690f7d-0a21-4296-b677-3c30f7cb878a-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.436704 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.551604 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.587514 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.603683 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:07:21 crc kubenswrapper[4767]: E0317 16:07:21.604728 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="ceilometer-central-agent" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.604754 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="ceilometer-central-agent" Mar 17 16:07:21 crc kubenswrapper[4767]: E0317 16:07:21.604768 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="ceilometer-notification-agent" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.604781 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="ceilometer-notification-agent" Mar 17 16:07:21 crc kubenswrapper[4767]: E0317 16:07:21.604813 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e690f7d-0a21-4296-b677-3c30f7cb878a" containerName="nova-scheduler-scheduler" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.604824 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e690f7d-0a21-4296-b677-3c30f7cb878a" containerName="nova-scheduler-scheduler" Mar 17 16:07:21 crc kubenswrapper[4767]: E0317 16:07:21.604854 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="sg-core" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.604863 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="sg-core" Mar 17 16:07:21 crc kubenswrapper[4767]: E0317 16:07:21.604886 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="proxy-httpd" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.604894 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="proxy-httpd" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.605315 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="proxy-httpd" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.605358 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e690f7d-0a21-4296-b677-3c30f7cb878a" containerName="nova-scheduler-scheduler" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.605377 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="sg-core" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.605394 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="ceilometer-notification-agent" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.605406 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" containerName="ceilometer-central-agent" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.606367 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-combined-ca-bundle\") pod \"c74426cc-3c82-4459-b9ab-ba39730f182a\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.606497 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn7kt\" (UniqueName: \"kubernetes.io/projected/c74426cc-3c82-4459-b9ab-ba39730f182a-kube-api-access-wn7kt\") pod \"c74426cc-3c82-4459-b9ab-ba39730f182a\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.606663 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-sg-core-conf-yaml\") pod \"c74426cc-3c82-4459-b9ab-ba39730f182a\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.606806 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-config-data\") pod \"c74426cc-3c82-4459-b9ab-ba39730f182a\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.606829 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-run-httpd\") pod \"c74426cc-3c82-4459-b9ab-ba39730f182a\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.606881 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-scripts\") pod \"c74426cc-3c82-4459-b9ab-ba39730f182a\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.606916 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-log-httpd\") pod \"c74426cc-3c82-4459-b9ab-ba39730f182a\" (UID: \"c74426cc-3c82-4459-b9ab-ba39730f182a\") " Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.607472 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.608279 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c74426cc-3c82-4459-b9ab-ba39730f182a" (UID: "c74426cc-3c82-4459-b9ab-ba39730f182a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.609493 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c74426cc-3c82-4459-b9ab-ba39730f182a" (UID: "c74426cc-3c82-4459-b9ab-ba39730f182a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.613879 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.616145 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-scripts" (OuterVolumeSpecName: "scripts") pod "c74426cc-3c82-4459-b9ab-ba39730f182a" (UID: "c74426cc-3c82-4459-b9ab-ba39730f182a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.617997 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.640639 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c74426cc-3c82-4459-b9ab-ba39730f182a-kube-api-access-wn7kt" (OuterVolumeSpecName: "kube-api-access-wn7kt") pod "c74426cc-3c82-4459-b9ab-ba39730f182a" (UID: "c74426cc-3c82-4459-b9ab-ba39730f182a"). InnerVolumeSpecName "kube-api-access-wn7kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.880841 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9074a89e-280b-4ba5-a680-f4a9084d4609-config-data\") pod \"nova-scheduler-0\" (UID: \"9074a89e-280b-4ba5-a680-f4a9084d4609\") " pod="openstack/nova-scheduler-0" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.881179 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brgw8\" (UniqueName: \"kubernetes.io/projected/9074a89e-280b-4ba5-a680-f4a9084d4609-kube-api-access-brgw8\") pod \"nova-scheduler-0\" (UID: \"9074a89e-280b-4ba5-a680-f4a9084d4609\") " pod="openstack/nova-scheduler-0" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.881236 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9074a89e-280b-4ba5-a680-f4a9084d4609-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9074a89e-280b-4ba5-a680-f4a9084d4609\") " pod="openstack/nova-scheduler-0" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.888011 4767 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.888418 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.888433 4767 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74426cc-3c82-4459-b9ab-ba39730f182a-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.888444 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn7kt\" (UniqueName: \"kubernetes.io/projected/c74426cc-3c82-4459-b9ab-ba39730f182a-kube-api-access-wn7kt\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:21 crc kubenswrapper[4767]: I0317 16:07:21.943281 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c74426cc-3c82-4459-b9ab-ba39730f182a" (UID: "c74426cc-3c82-4459-b9ab-ba39730f182a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.001974 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9074a89e-280b-4ba5-a680-f4a9084d4609-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9074a89e-280b-4ba5-a680-f4a9084d4609\") " pod="openstack/nova-scheduler-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.003694 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9074a89e-280b-4ba5-a680-f4a9084d4609-config-data\") pod \"nova-scheduler-0\" (UID: \"9074a89e-280b-4ba5-a680-f4a9084d4609\") " pod="openstack/nova-scheduler-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.003813 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brgw8\" (UniqueName: \"kubernetes.io/projected/9074a89e-280b-4ba5-a680-f4a9084d4609-kube-api-access-brgw8\") pod \"nova-scheduler-0\" (UID: \"9074a89e-280b-4ba5-a680-f4a9084d4609\") " pod="openstack/nova-scheduler-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.015728 4767 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.019360 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9074a89e-280b-4ba5-a680-f4a9084d4609-config-data\") pod \"nova-scheduler-0\" (UID: \"9074a89e-280b-4ba5-a680-f4a9084d4609\") " pod="openstack/nova-scheduler-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.029897 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9074a89e-280b-4ba5-a680-f4a9084d4609-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9074a89e-280b-4ba5-a680-f4a9084d4609\") " pod="openstack/nova-scheduler-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.035912 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brgw8\" (UniqueName: \"kubernetes.io/projected/9074a89e-280b-4ba5-a680-f4a9084d4609-kube-api-access-brgw8\") pod \"nova-scheduler-0\" (UID: \"9074a89e-280b-4ba5-a680-f4a9084d4609\") " pod="openstack/nova-scheduler-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.063287 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-config-data" (OuterVolumeSpecName: "config-data") pod "c74426cc-3c82-4459-b9ab-ba39730f182a" (UID: "c74426cc-3c82-4459-b9ab-ba39730f182a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.076201 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c74426cc-3c82-4459-b9ab-ba39730f182a" (UID: "c74426cc-3c82-4459-b9ab-ba39730f182a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.143531 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.143939 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74426cc-3c82-4459-b9ab-ba39730f182a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.221948 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.222155 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e3e3d43-9308-4bb7-8046-adf72dda25fe","Type":"ContainerStarted","Data":"68220034283daa22093b6de1679bbb27c500e46a58d9e9ca42a30b1e9a856604"} Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.222212 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e3e3d43-9308-4bb7-8046-adf72dda25fe","Type":"ContainerStarted","Data":"3ae3af5b71e3e924b44bbc4abedac6e5d23c25d2b728d078a99bb4197043856b"} Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.236019 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.236436 4767 generic.go:334] "Generic (PLEG): container finished" podID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerID="109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb" exitCode=137 Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.236856 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"f93b4e48-0fc9-4056-95c7-233caca8daf8","Type":"ContainerDied","Data":"109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb"} Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.236885 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"f93b4e48-0fc9-4056-95c7-233caca8daf8","Type":"ContainerDied","Data":"4f2bee65bec5fc21fef6a2036ea4a22d107fc422dcb79b0b1adcf52a58540c7f"} Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.236903 4767 scope.go:117] "RemoveContainer" containerID="109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.256016 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74426cc-3c82-4459-b9ab-ba39730f182a","Type":"ContainerDied","Data":"c3bc1821db2037b1eb67e35b95149b6081e5d93429037354563100af08e50c49"} Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.256278 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.263948 4767 generic.go:334] "Generic (PLEG): container finished" podID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerID="6e117d82f17da24ab40d63740750d5351b29e66934dbcc8001415c37592500dd" exitCode=0 Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.264006 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a81e7bf-b58a-4421-a981-4a69221f9a5f","Type":"ContainerDied","Data":"6e117d82f17da24ab40d63740750d5351b29e66934dbcc8001415c37592500dd"} Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.298513 4767 scope.go:117] "RemoveContainer" containerID="d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.349670 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-config-data\") pod \"f93b4e48-0fc9-4056-95c7-233caca8daf8\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.349851 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-scripts\") pod \"f93b4e48-0fc9-4056-95c7-233caca8daf8\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.349873 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnk8d\" (UniqueName: \"kubernetes.io/projected/f93b4e48-0fc9-4056-95c7-233caca8daf8-kube-api-access-wnk8d\") pod \"f93b4e48-0fc9-4056-95c7-233caca8daf8\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.349896 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-combined-ca-bundle\") pod \"f93b4e48-0fc9-4056-95c7-233caca8daf8\" (UID: \"f93b4e48-0fc9-4056-95c7-233caca8daf8\") " Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.357223 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.359227 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f93b4e48-0fc9-4056-95c7-233caca8daf8-kube-api-access-wnk8d" (OuterVolumeSpecName: "kube-api-access-wnk8d") pod "f93b4e48-0fc9-4056-95c7-233caca8daf8" (UID: "f93b4e48-0fc9-4056-95c7-233caca8daf8"). InnerVolumeSpecName "kube-api-access-wnk8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.371626 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-scripts" (OuterVolumeSpecName: "scripts") pod "f93b4e48-0fc9-4056-95c7-233caca8daf8" (UID: "f93b4e48-0fc9-4056-95c7-233caca8daf8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.398930 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.400938 4767 scope.go:117] "RemoveContainer" containerID="839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.428210 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:22 crc kubenswrapper[4767]: E0317 16:07:22.429292 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-evaluator" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.429314 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-evaluator" Mar 17 16:07:22 crc kubenswrapper[4767]: E0317 16:07:22.429328 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-notifier" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.429405 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-notifier" Mar 17 16:07:22 crc kubenswrapper[4767]: E0317 16:07:22.429508 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-api" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.429519 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-api" Mar 17 16:07:22 crc kubenswrapper[4767]: E0317 16:07:22.429579 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-listener" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.429585 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-listener" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.430034 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-listener" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.430058 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-evaluator" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.430080 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-api" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.430102 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" containerName="aodh-notifier" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.433701 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.449919 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.449927 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.462037 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnk8d\" (UniqueName: \"kubernetes.io/projected/f93b4e48-0fc9-4056-95c7-233caca8daf8-kube-api-access-wnk8d\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.462076 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.463375 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.471211 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.507192 4767 scope.go:117] "RemoveContainer" containerID="a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.570787 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f93b4e48-0fc9-4056-95c7-233caca8daf8" (UID: "f93b4e48-0fc9-4056-95c7-233caca8daf8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.572233 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv5td\" (UniqueName: \"kubernetes.io/projected/8a81e7bf-b58a-4421-a981-4a69221f9a5f-kube-api-access-jv5td\") pod \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.572328 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-config-data\") pod \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.572478 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-nova-metadata-tls-certs\") pod \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.572766 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-combined-ca-bundle\") pod \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.572903 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a81e7bf-b58a-4421-a981-4a69221f9a5f-logs\") pod \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\" (UID: \"8a81e7bf-b58a-4421-a981-4a69221f9a5f\") " Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.574278 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-scripts\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.574317 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-log-httpd\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.574427 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.574580 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsdfz\" (UniqueName: \"kubernetes.io/projected/bfee8a92-09c3-424e-bdb9-973c50a10322-kube-api-access-bsdfz\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.575388 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-config-data\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.575428 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-run-httpd\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.575470 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.575484 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a81e7bf-b58a-4421-a981-4a69221f9a5f-logs" (OuterVolumeSpecName: "logs") pod "8a81e7bf-b58a-4421-a981-4a69221f9a5f" (UID: "8a81e7bf-b58a-4421-a981-4a69221f9a5f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.575655 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.575679 4767 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a81e7bf-b58a-4421-a981-4a69221f9a5f-logs\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.577453 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a81e7bf-b58a-4421-a981-4a69221f9a5f-kube-api-access-jv5td" (OuterVolumeSpecName: "kube-api-access-jv5td") pod "8a81e7bf-b58a-4421-a981-4a69221f9a5f" (UID: "8a81e7bf-b58a-4421-a981-4a69221f9a5f"). InnerVolumeSpecName "kube-api-access-jv5td". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.590965 4767 scope.go:117] "RemoveContainer" containerID="109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb" Mar 17 16:07:22 crc kubenswrapper[4767]: E0317 16:07:22.591689 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb\": container with ID starting with 109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb not found: ID does not exist" containerID="109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.591744 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb"} err="failed to get container status \"109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb\": rpc error: code = NotFound desc = could not find container \"109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb\": container with ID starting with 109d1c2cec19524e4d619c63d244ef2914fa03ce8f7675074267418e5caaccfb not found: ID does not exist" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.591830 4767 scope.go:117] "RemoveContainer" containerID="d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd" Mar 17 16:07:22 crc kubenswrapper[4767]: E0317 16:07:22.592611 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd\": container with ID starting with d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd not found: ID does not exist" containerID="d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.592663 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd"} err="failed to get container status \"d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd\": rpc error: code = NotFound desc = could not find container \"d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd\": container with ID starting with d017d0cc984ff5b42e4e4aa15166f771a522aef121c48e16471a23b7c3d3f9bd not found: ID does not exist" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.592738 4767 scope.go:117] "RemoveContainer" containerID="839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3" Mar 17 16:07:22 crc kubenswrapper[4767]: E0317 16:07:22.597596 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3\": container with ID starting with 839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3 not found: ID does not exist" containerID="839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.597646 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3"} err="failed to get container status \"839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3\": rpc error: code = NotFound desc = could not find container \"839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3\": container with ID starting with 839d0686c5b5fa323c1bf9c6a503aaf6273b1482c3187c39304c841588ae9ad3 not found: ID does not exist" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.597679 4767 scope.go:117] "RemoveContainer" containerID="a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280" Mar 17 16:07:22 crc kubenswrapper[4767]: E0317 16:07:22.598130 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280\": container with ID starting with a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280 not found: ID does not exist" containerID="a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.598154 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280"} err="failed to get container status \"a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280\": rpc error: code = NotFound desc = could not find container \"a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280\": container with ID starting with a04c1d20ac79118e110cd4084e6b0d381ec7ee8dd729014188f9bc8f3ad6b280 not found: ID does not exist" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.598292 4767 scope.go:117] "RemoveContainer" containerID="46765c1bf5a2acbafd3d6c783f2fdb5d3eaceea216a84bad457961ede06ce025" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.624457 4767 scope.go:117] "RemoveContainer" containerID="b3e6754c1194ec22767e5bdf230d7f6e7b1e92fdc39e15b6f061f7ad326e0701" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.644211 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-config-data" (OuterVolumeSpecName: "config-data") pod "8a81e7bf-b58a-4421-a981-4a69221f9a5f" (UID: "8a81e7bf-b58a-4421-a981-4a69221f9a5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.651505 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a81e7bf-b58a-4421-a981-4a69221f9a5f" (UID: "8a81e7bf-b58a-4421-a981-4a69221f9a5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.668797 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-config-data" (OuterVolumeSpecName: "config-data") pod "f93b4e48-0fc9-4056-95c7-233caca8daf8" (UID: "f93b4e48-0fc9-4056-95c7-233caca8daf8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.678161 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-scripts\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.678746 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-log-httpd\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.678837 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.678957 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsdfz\" (UniqueName: \"kubernetes.io/projected/bfee8a92-09c3-424e-bdb9-973c50a10322-kube-api-access-bsdfz\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.679140 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-run-httpd\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.679190 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-config-data\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.679217 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.679289 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.679281 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-log-httpd\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.679302 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.679356 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93b4e48-0fc9-4056-95c7-233caca8daf8-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.679378 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv5td\" (UniqueName: \"kubernetes.io/projected/8a81e7bf-b58a-4421-a981-4a69221f9a5f-kube-api-access-jv5td\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.680837 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-run-httpd\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.686019 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.688655 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.690618 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-config-data\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.693424 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-scripts\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.732157 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsdfz\" (UniqueName: \"kubernetes.io/projected/bfee8a92-09c3-424e-bdb9-973c50a10322-kube-api-access-bsdfz\") pod \"ceilometer-0\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.734166 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8a81e7bf-b58a-4421-a981-4a69221f9a5f" (UID: "8a81e7bf-b58a-4421-a981-4a69221f9a5f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.786122 4767 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a81e7bf-b58a-4421-a981-4a69221f9a5f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.799996 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.859986 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 16:07:22 crc kubenswrapper[4767]: I0317 16:07:22.995338 4767 scope.go:117] "RemoveContainer" containerID="cc6095ba10d1bedc84ab4b3e688571375cf038de71aac3f583b5df9dda17fcbc" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.047418 4767 scope.go:117] "RemoveContainer" containerID="badd6c9149b8055fe6ed0453e3037e93998c97262d51784b281eb766d15262a9" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.281478 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9074a89e-280b-4ba5-a680-f4a9084d4609","Type":"ContainerStarted","Data":"7d8945ab52de07bf744e5baaba5e2e9a94cc952155aeee67c223e81bce444ced"} Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.281837 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9074a89e-280b-4ba5-a680-f4a9084d4609","Type":"ContainerStarted","Data":"334f364f71bfb752868ddb9617ef8f3418d360d8673d7686f8bcf33595f64dca"} Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.285164 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.305726 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a81e7bf-b58a-4421-a981-4a69221f9a5f","Type":"ContainerDied","Data":"636b97d2519b84768102c0a14a4f7d5b7415c99625dfef25fc09ef40eea89c9f"} Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.305745 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.305796 4767 scope.go:117] "RemoveContainer" containerID="6e117d82f17da24ab40d63740750d5351b29e66934dbcc8001415c37592500dd" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.310956 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.31093106 podStartE2EDuration="2.31093106s" podCreationTimestamp="2026-03-17 16:07:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:07:23.307146851 +0000 UTC m=+1834.720462928" watchObservedRunningTime="2026-03-17 16:07:23.31093106 +0000 UTC m=+1834.724247107" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.312927 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e3e3d43-9308-4bb7-8046-adf72dda25fe","Type":"ContainerStarted","Data":"5b46198130b736c26c94f5e13b7355800e22e052a9e1d216b7dded804491d513"} Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.344527 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.357218 4767 scope.go:117] "RemoveContainer" containerID="f3cf633b26e09018051f5fe48015ec062e4e687ac9f31c953ac6def5c9069f21" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.384345 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e690f7d-0a21-4296-b677-3c30f7cb878a" path="/var/lib/kubelet/pods/2e690f7d-0a21-4296-b677-3c30f7cb878a/volumes" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.395871 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c74426cc-3c82-4459-b9ab-ba39730f182a" path="/var/lib/kubelet/pods/c74426cc-3c82-4459-b9ab-ba39730f182a/volumes" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.408459 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.408521 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.408539 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.433259 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Mar 17 16:07:23 crc kubenswrapper[4767]: E0317 16:07:23.434097 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerName="nova-metadata-log" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.434114 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerName="nova-metadata-log" Mar 17 16:07:23 crc kubenswrapper[4767]: E0317 16:07:23.434131 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerName="nova-metadata-metadata" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.434137 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerName="nova-metadata-metadata" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.434478 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerName="nova-metadata-log" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.434530 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" containerName="nova-metadata-metadata" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.436958 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.441345 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-9wfxq" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.442061 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.442141 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.442244 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.445112 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.447451 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.465549 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.466298 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.466275683 podStartE2EDuration="3.466275683s" podCreationTimestamp="2026-03-17 16:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:07:23.398750871 +0000 UTC m=+1834.812066928" watchObservedRunningTime="2026-03-17 16:07:23.466275683 +0000 UTC m=+1834.879591730" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.477824 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.485161 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.485503 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.507029 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-internal-tls-certs\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.507324 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-public-tls-certs\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.507377 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-scripts\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.507537 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnv6j\" (UniqueName: \"kubernetes.io/projected/4ec07ded-d23f-411d-a86f-4d971155bf1f-kube-api-access-mnv6j\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.507584 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.507686 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-config-data\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.533088 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.556162 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.610889 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-internal-tls-certs\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.611030 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-public-tls-certs\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.611068 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c8f38fa-bebf-40e0-9414-33fca2068790-logs\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.611096 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-scripts\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.621038 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c8f38fa-bebf-40e0-9414-33fca2068790-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.624451 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk2dt\" (UniqueName: \"kubernetes.io/projected/5c8f38fa-bebf-40e0-9414-33fca2068790-kube-api-access-rk2dt\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.624607 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnv6j\" (UniqueName: \"kubernetes.io/projected/4ec07ded-d23f-411d-a86f-4d971155bf1f-kube-api-access-mnv6j\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.624654 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.624743 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c8f38fa-bebf-40e0-9414-33fca2068790-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.626634 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-config-data\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.627049 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c8f38fa-bebf-40e0-9414-33fca2068790-config-data\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.635640 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-scripts\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.635795 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.638206 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-config-data\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.648032 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnv6j\" (UniqueName: \"kubernetes.io/projected/4ec07ded-d23f-411d-a86f-4d971155bf1f-kube-api-access-mnv6j\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.651626 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-public-tls-certs\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.655676 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-internal-tls-certs\") pod \"aodh-0\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.729854 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c8f38fa-bebf-40e0-9414-33fca2068790-logs\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.730656 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c8f38fa-bebf-40e0-9414-33fca2068790-logs\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.730666 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c8f38fa-bebf-40e0-9414-33fca2068790-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.730886 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk2dt\" (UniqueName: \"kubernetes.io/projected/5c8f38fa-bebf-40e0-9414-33fca2068790-kube-api-access-rk2dt\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.731112 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c8f38fa-bebf-40e0-9414-33fca2068790-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.731553 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c8f38fa-bebf-40e0-9414-33fca2068790-config-data\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.735556 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c8f38fa-bebf-40e0-9414-33fca2068790-config-data\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.735562 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c8f38fa-bebf-40e0-9414-33fca2068790-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.736265 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c8f38fa-bebf-40e0-9414-33fca2068790-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.749304 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk2dt\" (UniqueName: \"kubernetes.io/projected/5c8f38fa-bebf-40e0-9414-33fca2068790-kube-api-access-rk2dt\") pod \"nova-metadata-0\" (UID: \"5c8f38fa-bebf-40e0-9414-33fca2068790\") " pod="openstack/nova-metadata-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.799420 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 17 16:07:23 crc kubenswrapper[4767]: I0317 16:07:23.829809 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 16:07:24 crc kubenswrapper[4767]: I0317 16:07:24.356464 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfee8a92-09c3-424e-bdb9-973c50a10322","Type":"ContainerStarted","Data":"ca58a60067190c73fc0aa897136216d60546fbb2dc5d45adee71adac15fe35c2"} Mar 17 16:07:24 crc kubenswrapper[4767]: I0317 16:07:24.432413 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 17 16:07:24 crc kubenswrapper[4767]: W0317 16:07:24.743331 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c8f38fa_bebf_40e0_9414_33fca2068790.slice/crio-5fbf2a4e17321d84933a294010c59a843cf92a157995de32a2654647e8009dc6 WatchSource:0}: Error finding container 5fbf2a4e17321d84933a294010c59a843cf92a157995de32a2654647e8009dc6: Status 404 returned error can't find the container with id 5fbf2a4e17321d84933a294010c59a843cf92a157995de32a2654647e8009dc6 Mar 17 16:07:24 crc kubenswrapper[4767]: I0317 16:07:24.745894 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 16:07:25 crc kubenswrapper[4767]: I0317 16:07:25.388248 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a81e7bf-b58a-4421-a981-4a69221f9a5f" path="/var/lib/kubelet/pods/8a81e7bf-b58a-4421-a981-4a69221f9a5f/volumes" Mar 17 16:07:25 crc kubenswrapper[4767]: I0317 16:07:25.391425 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f93b4e48-0fc9-4056-95c7-233caca8daf8" path="/var/lib/kubelet/pods/f93b4e48-0fc9-4056-95c7-233caca8daf8/volumes" Mar 17 16:07:25 crc kubenswrapper[4767]: I0317 16:07:25.414232 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfee8a92-09c3-424e-bdb9-973c50a10322","Type":"ContainerStarted","Data":"16ed2ddd83e90b96b95a234eaf96bcd80511e23d2a37782421f0d5b119b47c46"} Mar 17 16:07:25 crc kubenswrapper[4767]: I0317 16:07:25.416127 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c8f38fa-bebf-40e0-9414-33fca2068790","Type":"ContainerStarted","Data":"5fc5062d27bfb033c9f3e4b4e8867eb848f4e0e47c307b3cf3398ac49bde3ea5"} Mar 17 16:07:25 crc kubenswrapper[4767]: I0317 16:07:25.416209 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c8f38fa-bebf-40e0-9414-33fca2068790","Type":"ContainerStarted","Data":"5fbf2a4e17321d84933a294010c59a843cf92a157995de32a2654647e8009dc6"} Mar 17 16:07:25 crc kubenswrapper[4767]: I0317 16:07:25.417762 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4ec07ded-d23f-411d-a86f-4d971155bf1f","Type":"ContainerStarted","Data":"daede6d5b050635880d872508b8da6238cef7b3d1998772b33488a3e45ad907a"} Mar 17 16:07:26 crc kubenswrapper[4767]: I0317 16:07:26.445562 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4ec07ded-d23f-411d-a86f-4d971155bf1f","Type":"ContainerStarted","Data":"aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c"} Mar 17 16:07:26 crc kubenswrapper[4767]: I0317 16:07:26.446201 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4ec07ded-d23f-411d-a86f-4d971155bf1f","Type":"ContainerStarted","Data":"9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7"} Mar 17 16:07:26 crc kubenswrapper[4767]: I0317 16:07:26.453784 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfee8a92-09c3-424e-bdb9-973c50a10322","Type":"ContainerStarted","Data":"c2ba3abf835e8e32b378663c7dbe6060b1e565bed7bbcdb5aee104ef1acaec45"} Mar 17 16:07:26 crc kubenswrapper[4767]: I0317 16:07:26.453845 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfee8a92-09c3-424e-bdb9-973c50a10322","Type":"ContainerStarted","Data":"d4a1f434e7c4a4ce983ca40097c549cebd416bd03f084596b19e3c868a329f18"} Mar 17 16:07:26 crc kubenswrapper[4767]: I0317 16:07:26.456675 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c8f38fa-bebf-40e0-9414-33fca2068790","Type":"ContainerStarted","Data":"0bcad43eb474c16d5fb0bdcfb9e29d584a21153b7dcf3ff15ed507c6b5b414ba"} Mar 17 16:07:26 crc kubenswrapper[4767]: I0317 16:07:26.480778 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.480754241 podStartE2EDuration="3.480754241s" podCreationTimestamp="2026-03-17 16:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:07:26.478777725 +0000 UTC m=+1837.892093802" watchObservedRunningTime="2026-03-17 16:07:26.480754241 +0000 UTC m=+1837.894070288" Mar 17 16:07:27 crc kubenswrapper[4767]: I0317 16:07:27.222717 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 17 16:07:27 crc kubenswrapper[4767]: I0317 16:07:27.472312 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4ec07ded-d23f-411d-a86f-4d971155bf1f","Type":"ContainerStarted","Data":"c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe"} Mar 17 16:07:28 crc kubenswrapper[4767]: I0317 16:07:28.507846 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4ec07ded-d23f-411d-a86f-4d971155bf1f","Type":"ContainerStarted","Data":"5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775"} Mar 17 16:07:28 crc kubenswrapper[4767]: I0317 16:07:28.561674 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.418799949 podStartE2EDuration="5.561647979s" podCreationTimestamp="2026-03-17 16:07:23 +0000 UTC" firstStartedPulling="2026-03-17 16:07:24.444464471 +0000 UTC m=+1835.857780518" lastFinishedPulling="2026-03-17 16:07:27.587312501 +0000 UTC m=+1839.000628548" observedRunningTime="2026-03-17 16:07:28.529128649 +0000 UTC m=+1839.942444696" watchObservedRunningTime="2026-03-17 16:07:28.561647979 +0000 UTC m=+1839.974964026" Mar 17 16:07:29 crc kubenswrapper[4767]: I0317 16:07:29.523980 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfee8a92-09c3-424e-bdb9-973c50a10322","Type":"ContainerStarted","Data":"ed4d3c6c450409bf7c4b5a8866365f5b5d9eaf77773bfe9a48bc9d2b4361696f"} Mar 17 16:07:29 crc kubenswrapper[4767]: I0317 16:07:29.555804 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.036561105 podStartE2EDuration="7.555772122s" podCreationTimestamp="2026-03-17 16:07:22 +0000 UTC" firstStartedPulling="2026-03-17 16:07:23.512593977 +0000 UTC m=+1834.925910024" lastFinishedPulling="2026-03-17 16:07:28.031804994 +0000 UTC m=+1839.445121041" observedRunningTime="2026-03-17 16:07:29.553754814 +0000 UTC m=+1840.967070871" watchObservedRunningTime="2026-03-17 16:07:29.555772122 +0000 UTC m=+1840.969088169" Mar 17 16:07:30 crc kubenswrapper[4767]: I0317 16:07:30.540676 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 16:07:30 crc kubenswrapper[4767]: I0317 16:07:30.576847 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 16:07:30 crc kubenswrapper[4767]: I0317 16:07:30.578830 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 16:07:31 crc kubenswrapper[4767]: I0317 16:07:31.594454 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2e3e3d43-9308-4bb7-8046-adf72dda25fe" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.13:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 16:07:31 crc kubenswrapper[4767]: I0317 16:07:31.595196 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2e3e3d43-9308-4bb7-8046-adf72dda25fe" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.13:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 16:07:32 crc kubenswrapper[4767]: I0317 16:07:32.222858 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 17 16:07:32 crc kubenswrapper[4767]: I0317 16:07:32.273372 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 17 16:07:32 crc kubenswrapper[4767]: I0317 16:07:32.354624 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:07:32 crc kubenswrapper[4767]: E0317 16:07:32.354962 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:07:32 crc kubenswrapper[4767]: I0317 16:07:32.599626 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 17 16:07:33 crc kubenswrapper[4767]: I0317 16:07:33.830811 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 17 16:07:33 crc kubenswrapper[4767]: I0317 16:07:33.831359 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 17 16:07:34 crc kubenswrapper[4767]: I0317 16:07:34.848598 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5c8f38fa-bebf-40e0-9414-33fca2068790" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.17:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 16:07:34 crc kubenswrapper[4767]: I0317 16:07:34.848600 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5c8f38fa-bebf-40e0-9414-33fca2068790" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.17:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 16:07:38 crc kubenswrapper[4767]: I0317 16:07:38.576284 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 17 16:07:38 crc kubenswrapper[4767]: I0317 16:07:38.576926 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 17 16:07:40 crc kubenswrapper[4767]: I0317 16:07:40.585566 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 17 16:07:40 crc kubenswrapper[4767]: I0317 16:07:40.586294 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 17 16:07:40 crc kubenswrapper[4767]: I0317 16:07:40.594322 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 17 16:07:40 crc kubenswrapper[4767]: I0317 16:07:40.602136 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 17 16:07:41 crc kubenswrapper[4767]: I0317 16:07:41.831432 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 17 16:07:41 crc kubenswrapper[4767]: I0317 16:07:41.831733 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 17 16:07:43 crc kubenswrapper[4767]: I0317 16:07:43.836595 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 17 16:07:43 crc kubenswrapper[4767]: I0317 16:07:43.840717 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 17 16:07:43 crc kubenswrapper[4767]: I0317 16:07:43.845155 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 17 16:07:44 crc kubenswrapper[4767]: I0317 16:07:44.727133 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 17 16:07:46 crc kubenswrapper[4767]: I0317 16:07:46.355707 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:07:46 crc kubenswrapper[4767]: E0317 16:07:46.356399 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.067544 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lc99p"] Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.084414 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.092375 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lc99p"] Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.479766 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk9t4\" (UniqueName: \"kubernetes.io/projected/1dea3dfa-5821-4999-99f8-7c17d9ec7515-kube-api-access-pk9t4\") pod \"redhat-operators-lc99p\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.481088 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-catalog-content\") pod \"redhat-operators-lc99p\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.486121 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-utilities\") pod \"redhat-operators-lc99p\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.590237 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-utilities\") pod \"redhat-operators-lc99p\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.590323 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk9t4\" (UniqueName: \"kubernetes.io/projected/1dea3dfa-5821-4999-99f8-7c17d9ec7515-kube-api-access-pk9t4\") pod \"redhat-operators-lc99p\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.590457 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-catalog-content\") pod \"redhat-operators-lc99p\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.591833 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-utilities\") pod \"redhat-operators-lc99p\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.592653 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-catalog-content\") pod \"redhat-operators-lc99p\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.613599 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk9t4\" (UniqueName: \"kubernetes.io/projected/1dea3dfa-5821-4999-99f8-7c17d9ec7515-kube-api-access-pk9t4\") pod \"redhat-operators-lc99p\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:48 crc kubenswrapper[4767]: I0317 16:07:48.717080 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:49 crc kubenswrapper[4767]: I0317 16:07:49.317650 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lc99p"] Mar 17 16:07:49 crc kubenswrapper[4767]: I0317 16:07:49.785527 4767 generic.go:334] "Generic (PLEG): container finished" podID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerID="adf8ca6066c5497a961566785101b0324f1fe442e2ebcc3db1c6e8d5b3b13044" exitCode=0 Mar 17 16:07:49 crc kubenswrapper[4767]: I0317 16:07:49.785625 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc99p" event={"ID":"1dea3dfa-5821-4999-99f8-7c17d9ec7515","Type":"ContainerDied","Data":"adf8ca6066c5497a961566785101b0324f1fe442e2ebcc3db1c6e8d5b3b13044"} Mar 17 16:07:49 crc kubenswrapper[4767]: I0317 16:07:49.786084 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc99p" event={"ID":"1dea3dfa-5821-4999-99f8-7c17d9ec7515","Type":"ContainerStarted","Data":"84bbb39b8f9c69ebbb6ee248a2254287c39668677cd66596baf7a934229d8c89"} Mar 17 16:07:51 crc kubenswrapper[4767]: I0317 16:07:51.834441 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc99p" event={"ID":"1dea3dfa-5821-4999-99f8-7c17d9ec7515","Type":"ContainerStarted","Data":"6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a"} Mar 17 16:07:52 crc kubenswrapper[4767]: I0317 16:07:52.810396 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 17 16:07:56 crc kubenswrapper[4767]: I0317 16:07:56.974196 4767 generic.go:334] "Generic (PLEG): container finished" podID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerID="6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a" exitCode=0 Mar 17 16:07:56 crc kubenswrapper[4767]: I0317 16:07:56.974326 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc99p" event={"ID":"1dea3dfa-5821-4999-99f8-7c17d9ec7515","Type":"ContainerDied","Data":"6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a"} Mar 17 16:07:57 crc kubenswrapper[4767]: I0317 16:07:57.917546 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 16:07:57 crc kubenswrapper[4767]: I0317 16:07:57.918159 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="232c4872-a09a-4be2-a7d2-c3c7232bae8a" containerName="kube-state-metrics" containerID="cri-o://74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e" gracePeriod=30 Mar 17 16:07:57 crc kubenswrapper[4767]: I0317 16:07:57.991200 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc99p" event={"ID":"1dea3dfa-5821-4999-99f8-7c17d9ec7515","Type":"ContainerStarted","Data":"0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5"} Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.020954 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lc99p" podStartSLOduration=2.417602653 podStartE2EDuration="10.020909999s" podCreationTimestamp="2026-03-17 16:07:48 +0000 UTC" firstStartedPulling="2026-03-17 16:07:49.789164399 +0000 UTC m=+1861.202480446" lastFinishedPulling="2026-03-17 16:07:57.392471745 +0000 UTC m=+1868.805787792" observedRunningTime="2026-03-17 16:07:58.015567946 +0000 UTC m=+1869.428884013" watchObservedRunningTime="2026-03-17 16:07:58.020909999 +0000 UTC m=+1869.434226056" Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.105436 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.106170 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="bf461aeb-5692-40f0-be55-16e509064dc6" containerName="mysqld-exporter" containerID="cri-o://1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8" gracePeriod=30 Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.597992 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.717523 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.717840 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.744249 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knd9k\" (UniqueName: \"kubernetes.io/projected/232c4872-a09a-4be2-a7d2-c3c7232bae8a-kube-api-access-knd9k\") pod \"232c4872-a09a-4be2-a7d2-c3c7232bae8a\" (UID: \"232c4872-a09a-4be2-a7d2-c3c7232bae8a\") " Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.756686 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/232c4872-a09a-4be2-a7d2-c3c7232bae8a-kube-api-access-knd9k" (OuterVolumeSpecName: "kube-api-access-knd9k") pod "232c4872-a09a-4be2-a7d2-c3c7232bae8a" (UID: "232c4872-a09a-4be2-a7d2-c3c7232bae8a"). InnerVolumeSpecName "kube-api-access-knd9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.859918 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.869687 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knd9k\" (UniqueName: \"kubernetes.io/projected/232c4872-a09a-4be2-a7d2-c3c7232bae8a-kube-api-access-knd9k\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.972508 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-config-data\") pod \"bf461aeb-5692-40f0-be55-16e509064dc6\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.972709 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-combined-ca-bundle\") pod \"bf461aeb-5692-40f0-be55-16e509064dc6\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " Mar 17 16:07:58 crc kubenswrapper[4767]: I0317 16:07:58.972803 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dr5v\" (UniqueName: \"kubernetes.io/projected/bf461aeb-5692-40f0-be55-16e509064dc6-kube-api-access-4dr5v\") pod \"bf461aeb-5692-40f0-be55-16e509064dc6\" (UID: \"bf461aeb-5692-40f0-be55-16e509064dc6\") " Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.004754 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf461aeb-5692-40f0-be55-16e509064dc6-kube-api-access-4dr5v" (OuterVolumeSpecName: "kube-api-access-4dr5v") pod "bf461aeb-5692-40f0-be55-16e509064dc6" (UID: "bf461aeb-5692-40f0-be55-16e509064dc6"). InnerVolumeSpecName "kube-api-access-4dr5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.030635 4767 generic.go:334] "Generic (PLEG): container finished" podID="bf461aeb-5692-40f0-be55-16e509064dc6" containerID="1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8" exitCode=2 Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.030747 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"bf461aeb-5692-40f0-be55-16e509064dc6","Type":"ContainerDied","Data":"1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8"} Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.030908 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"bf461aeb-5692-40f0-be55-16e509064dc6","Type":"ContainerDied","Data":"386ce5f130b75578e149d4e9df0c9e87531af50c61593416f9699a7ea301ef8b"} Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.030931 4767 scope.go:117] "RemoveContainer" containerID="1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.031105 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.050834 4767 generic.go:334] "Generic (PLEG): container finished" podID="232c4872-a09a-4be2-a7d2-c3c7232bae8a" containerID="74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e" exitCode=2 Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.051953 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.052557 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"232c4872-a09a-4be2-a7d2-c3c7232bae8a","Type":"ContainerDied","Data":"74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e"} Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.052600 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"232c4872-a09a-4be2-a7d2-c3c7232bae8a","Type":"ContainerDied","Data":"b6284e8bcfda5776b5b64b781869d4c02c978fce8f3fc85c85d92be08c33d037"} Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.064700 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf461aeb-5692-40f0-be55-16e509064dc6" (UID: "bf461aeb-5692-40f0-be55-16e509064dc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.076450 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.076507 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dr5v\" (UniqueName: \"kubernetes.io/projected/bf461aeb-5692-40f0-be55-16e509064dc6-kube-api-access-4dr5v\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.100384 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-config-data" (OuterVolumeSpecName: "config-data") pod "bf461aeb-5692-40f0-be55-16e509064dc6" (UID: "bf461aeb-5692-40f0-be55-16e509064dc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.156135 4767 scope.go:117] "RemoveContainer" containerID="1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8" Mar 17 16:07:59 crc kubenswrapper[4767]: E0317 16:07:59.156862 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8\": container with ID starting with 1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8 not found: ID does not exist" containerID="1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.156910 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8"} err="failed to get container status \"1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8\": rpc error: code = NotFound desc = could not find container \"1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8\": container with ID starting with 1b71340a1d39dab1b46ff3c9368415fc185d36d7525d2b7708d9aff844a595c8 not found: ID does not exist" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.156942 4767 scope.go:117] "RemoveContainer" containerID="74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.163135 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.181633 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf461aeb-5692-40f0-be55-16e509064dc6-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:07:59 crc kubenswrapper[4767]: E0317 16:07:59.191804 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod232c4872_a09a_4be2_a7d2_c3c7232bae8a.slice/crio-b6284e8bcfda5776b5b64b781869d4c02c978fce8f3fc85c85d92be08c33d037\": RecentStats: unable to find data in memory cache]" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.204510 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.213794 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 16:07:59 crc kubenswrapper[4767]: E0317 16:07:59.214685 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232c4872-a09a-4be2-a7d2-c3c7232bae8a" containerName="kube-state-metrics" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.214711 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="232c4872-a09a-4be2-a7d2-c3c7232bae8a" containerName="kube-state-metrics" Mar 17 16:07:59 crc kubenswrapper[4767]: E0317 16:07:59.214748 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf461aeb-5692-40f0-be55-16e509064dc6" containerName="mysqld-exporter" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.214755 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf461aeb-5692-40f0-be55-16e509064dc6" containerName="mysqld-exporter" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.215036 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="232c4872-a09a-4be2-a7d2-c3c7232bae8a" containerName="kube-state-metrics" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.215078 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf461aeb-5692-40f0-be55-16e509064dc6" containerName="mysqld-exporter" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.216595 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.219705 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.220505 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.220500 4767 scope.go:117] "RemoveContainer" containerID="74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e" Mar 17 16:07:59 crc kubenswrapper[4767]: E0317 16:07:59.221943 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e\": container with ID starting with 74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e not found: ID does not exist" containerID="74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.221987 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e"} err="failed to get container status \"74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e\": rpc error: code = NotFound desc = could not find container \"74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e\": container with ID starting with 74e7f22ab7ed0764b7b354262300d572b3418268274c08f8d3bc233c338aa83e not found: ID does not exist" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.237135 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.386585 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="232c4872-a09a-4be2-a7d2-c3c7232bae8a" path="/var/lib/kubelet/pods/232c4872-a09a-4be2-a7d2-c3c7232bae8a/volumes" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.393365 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc5b680-771a-4719-81d6-08efe998c1de-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.393543 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdc5b680-771a-4719-81d6-08efe998c1de-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.393727 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bdc5b680-771a-4719-81d6-08efe998c1de-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.393871 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-784hm\" (UniqueName: \"kubernetes.io/projected/bdc5b680-771a-4719-81d6-08efe998c1de-kube-api-access-784hm\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.434282 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.460277 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.497248 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.500407 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.502920 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc5b680-771a-4719-81d6-08efe998c1de-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.503018 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdc5b680-771a-4719-81d6-08efe998c1de-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.503135 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bdc5b680-771a-4719-81d6-08efe998c1de-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.503285 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-784hm\" (UniqueName: \"kubernetes.io/projected/bdc5b680-771a-4719-81d6-08efe998c1de-kube-api-access-784hm\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.509805 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.509953 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.512112 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bdc5b680-771a-4719-81d6-08efe998c1de-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.514089 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdc5b680-771a-4719-81d6-08efe998c1de-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.524211 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc5b680-771a-4719-81d6-08efe998c1de-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.526163 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.535982 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-784hm\" (UniqueName: \"kubernetes.io/projected/bdc5b680-771a-4719-81d6-08efe998c1de-kube-api-access-784hm\") pod \"kube-state-metrics-0\" (UID: \"bdc5b680-771a-4719-81d6-08efe998c1de\") " pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.546142 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.729026 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20e80bf6-9aac-405d-b8f6-3435738f816f-config-data\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.729421 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20e80bf6-9aac-405d-b8f6-3435738f816f-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.729516 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqshm\" (UniqueName: \"kubernetes.io/projected/20e80bf6-9aac-405d-b8f6-3435738f816f-kube-api-access-dqshm\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.729583 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/20e80bf6-9aac-405d-b8f6-3435738f816f-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.837799 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lc99p" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="registry-server" probeResult="failure" output=< Mar 17 16:07:59 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:07:59 crc kubenswrapper[4767]: > Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.860160 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20e80bf6-9aac-405d-b8f6-3435738f816f-config-data\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.860269 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20e80bf6-9aac-405d-b8f6-3435738f816f-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.860452 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqshm\" (UniqueName: \"kubernetes.io/projected/20e80bf6-9aac-405d-b8f6-3435738f816f-kube-api-access-dqshm\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.860629 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/20e80bf6-9aac-405d-b8f6-3435738f816f-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.870015 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20e80bf6-9aac-405d-b8f6-3435738f816f-config-data\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.880427 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/20e80bf6-9aac-405d-b8f6-3435738f816f-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.880660 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20e80bf6-9aac-405d-b8f6-3435738f816f-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:07:59 crc kubenswrapper[4767]: I0317 16:07:59.889616 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqshm\" (UniqueName: \"kubernetes.io/projected/20e80bf6-9aac-405d-b8f6-3435738f816f-kube-api-access-dqshm\") pod \"mysqld-exporter-0\" (UID: \"20e80bf6-9aac-405d-b8f6-3435738f816f\") " pod="openstack/mysqld-exporter-0" Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.103006 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.160154 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562728-g6ss4"] Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.163501 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562728-g6ss4" Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.167041 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.167154 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.167344 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.172785 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfzvr\" (UniqueName: \"kubernetes.io/projected/68a9bbf0-4d9d-4760-9aa7-3afc7aa20123-kube-api-access-vfzvr\") pod \"auto-csr-approver-29562728-g6ss4\" (UID: \"68a9bbf0-4d9d-4760-9aa7-3afc7aa20123\") " pod="openshift-infra/auto-csr-approver-29562728-g6ss4" Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.187008 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562728-g6ss4"] Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.280922 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfzvr\" (UniqueName: \"kubernetes.io/projected/68a9bbf0-4d9d-4760-9aa7-3afc7aa20123-kube-api-access-vfzvr\") pod \"auto-csr-approver-29562728-g6ss4\" (UID: \"68a9bbf0-4d9d-4760-9aa7-3afc7aa20123\") " pod="openshift-infra/auto-csr-approver-29562728-g6ss4" Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.303363 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 16:08:00 crc kubenswrapper[4767]: W0317 16:08:00.311438 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdc5b680_771a_4719_81d6_08efe998c1de.slice/crio-e05b11095a58de43e6bf99590bcf25808e9f93593c6cfe69b811a407e28829eb WatchSource:0}: Error finding container e05b11095a58de43e6bf99590bcf25808e9f93593c6cfe69b811a407e28829eb: Status 404 returned error can't find the container with id e05b11095a58de43e6bf99590bcf25808e9f93593c6cfe69b811a407e28829eb Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.312918 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfzvr\" (UniqueName: \"kubernetes.io/projected/68a9bbf0-4d9d-4760-9aa7-3afc7aa20123-kube-api-access-vfzvr\") pod \"auto-csr-approver-29562728-g6ss4\" (UID: \"68a9bbf0-4d9d-4760-9aa7-3afc7aa20123\") " pod="openshift-infra/auto-csr-approver-29562728-g6ss4" Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.500608 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562728-g6ss4" Mar 17 16:08:00 crc kubenswrapper[4767]: I0317 16:08:00.788045 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 17 16:08:00 crc kubenswrapper[4767]: W0317 16:08:00.843365 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20e80bf6_9aac_405d_b8f6_3435738f816f.slice/crio-13723e552b6b15c820aba6c8e237dfbc94966d3d1e47c101144a0faa41d5fcd1 WatchSource:0}: Error finding container 13723e552b6b15c820aba6c8e237dfbc94966d3d1e47c101144a0faa41d5fcd1: Status 404 returned error can't find the container with id 13723e552b6b15c820aba6c8e237dfbc94966d3d1e47c101144a0faa41d5fcd1 Mar 17 16:08:01 crc kubenswrapper[4767]: W0317 16:08:01.073938 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68a9bbf0_4d9d_4760_9aa7_3afc7aa20123.slice/crio-8c6a071f702fa0daf4e81323d7b5a7f2021b4daf1d9581d73200fa86e5e2aa78 WatchSource:0}: Error finding container 8c6a071f702fa0daf4e81323d7b5a7f2021b4daf1d9581d73200fa86e5e2aa78: Status 404 returned error can't find the container with id 8c6a071f702fa0daf4e81323d7b5a7f2021b4daf1d9581d73200fa86e5e2aa78 Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.075443 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562728-g6ss4"] Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.098155 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"20e80bf6-9aac-405d-b8f6-3435738f816f","Type":"ContainerStarted","Data":"13723e552b6b15c820aba6c8e237dfbc94966d3d1e47c101144a0faa41d5fcd1"} Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.100713 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562728-g6ss4" event={"ID":"68a9bbf0-4d9d-4760-9aa7-3afc7aa20123","Type":"ContainerStarted","Data":"8c6a071f702fa0daf4e81323d7b5a7f2021b4daf1d9581d73200fa86e5e2aa78"} Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.103713 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bdc5b680-771a-4719-81d6-08efe998c1de","Type":"ContainerStarted","Data":"e05b11095a58de43e6bf99590bcf25808e9f93593c6cfe69b811a407e28829eb"} Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.103874 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.126917 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.697889985 podStartE2EDuration="2.126889815s" podCreationTimestamp="2026-03-17 16:07:59 +0000 UTC" firstStartedPulling="2026-03-17 16:08:00.318033751 +0000 UTC m=+1871.731349798" lastFinishedPulling="2026-03-17 16:08:00.747033571 +0000 UTC m=+1872.160349628" observedRunningTime="2026-03-17 16:08:01.125530526 +0000 UTC m=+1872.538846593" watchObservedRunningTime="2026-03-17 16:08:01.126889815 +0000 UTC m=+1872.540205872" Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.279738 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.280141 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="ceilometer-central-agent" containerID="cri-o://16ed2ddd83e90b96b95a234eaf96bcd80511e23d2a37782421f0d5b119b47c46" gracePeriod=30 Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.280231 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="proxy-httpd" containerID="cri-o://ed4d3c6c450409bf7c4b5a8866365f5b5d9eaf77773bfe9a48bc9d2b4361696f" gracePeriod=30 Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.280348 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="sg-core" containerID="cri-o://c2ba3abf835e8e32b378663c7dbe6060b1e565bed7bbcdb5aee104ef1acaec45" gracePeriod=30 Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.280428 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="ceilometer-notification-agent" containerID="cri-o://d4a1f434e7c4a4ce983ca40097c549cebd416bd03f084596b19e3c868a329f18" gracePeriod=30 Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.354976 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:08:01 crc kubenswrapper[4767]: E0317 16:08:01.355454 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:08:01 crc kubenswrapper[4767]: I0317 16:08:01.368581 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf461aeb-5692-40f0-be55-16e509064dc6" path="/var/lib/kubelet/pods/bf461aeb-5692-40f0-be55-16e509064dc6/volumes" Mar 17 16:08:02 crc kubenswrapper[4767]: I0317 16:08:02.123383 4767 generic.go:334] "Generic (PLEG): container finished" podID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerID="ed4d3c6c450409bf7c4b5a8866365f5b5d9eaf77773bfe9a48bc9d2b4361696f" exitCode=0 Mar 17 16:08:02 crc kubenswrapper[4767]: I0317 16:08:02.124051 4767 generic.go:334] "Generic (PLEG): container finished" podID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerID="c2ba3abf835e8e32b378663c7dbe6060b1e565bed7bbcdb5aee104ef1acaec45" exitCode=2 Mar 17 16:08:02 crc kubenswrapper[4767]: I0317 16:08:02.123482 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfee8a92-09c3-424e-bdb9-973c50a10322","Type":"ContainerDied","Data":"ed4d3c6c450409bf7c4b5a8866365f5b5d9eaf77773bfe9a48bc9d2b4361696f"} Mar 17 16:08:02 crc kubenswrapper[4767]: I0317 16:08:02.124155 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfee8a92-09c3-424e-bdb9-973c50a10322","Type":"ContainerDied","Data":"c2ba3abf835e8e32b378663c7dbe6060b1e565bed7bbcdb5aee104ef1acaec45"} Mar 17 16:08:02 crc kubenswrapper[4767]: I0317 16:08:02.124067 4767 generic.go:334] "Generic (PLEG): container finished" podID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerID="16ed2ddd83e90b96b95a234eaf96bcd80511e23d2a37782421f0d5b119b47c46" exitCode=0 Mar 17 16:08:02 crc kubenswrapper[4767]: I0317 16:08:02.124197 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfee8a92-09c3-424e-bdb9-973c50a10322","Type":"ContainerDied","Data":"16ed2ddd83e90b96b95a234eaf96bcd80511e23d2a37782421f0d5b119b47c46"} Mar 17 16:08:02 crc kubenswrapper[4767]: I0317 16:08:02.127121 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"20e80bf6-9aac-405d-b8f6-3435738f816f","Type":"ContainerStarted","Data":"6c01b87d7d13c073d4d4a7c5bd3ddaa5304e3d87ab5161b2af053dded9bd9f23"} Mar 17 16:08:02 crc kubenswrapper[4767]: I0317 16:08:02.129647 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bdc5b680-771a-4719-81d6-08efe998c1de","Type":"ContainerStarted","Data":"25c5824facc5159136ab627f7d2fd021867f75316e9ee1af1db1627aac9d91a7"} Mar 17 16:08:02 crc kubenswrapper[4767]: I0317 16:08:02.160306 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.622939163 podStartE2EDuration="3.160273571s" podCreationTimestamp="2026-03-17 16:07:59 +0000 UTC" firstStartedPulling="2026-03-17 16:08:00.857557922 +0000 UTC m=+1872.270873969" lastFinishedPulling="2026-03-17 16:08:01.39489233 +0000 UTC m=+1872.808208377" observedRunningTime="2026-03-17 16:08:02.144687056 +0000 UTC m=+1873.558003103" watchObservedRunningTime="2026-03-17 16:08:02.160273571 +0000 UTC m=+1873.573589618" Mar 17 16:08:03 crc kubenswrapper[4767]: I0317 16:08:03.154243 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562728-g6ss4" event={"ID":"68a9bbf0-4d9d-4760-9aa7-3afc7aa20123","Type":"ContainerStarted","Data":"71cee4369dbc400c2680a693359d1227aa47dda82465b728603476ef258ef922"} Mar 17 16:08:03 crc kubenswrapper[4767]: I0317 16:08:03.224874 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562728-g6ss4" podStartSLOduration=2.008337175 podStartE2EDuration="3.22485223s" podCreationTimestamp="2026-03-17 16:08:00 +0000 UTC" firstStartedPulling="2026-03-17 16:08:01.079000625 +0000 UTC m=+1872.492316672" lastFinishedPulling="2026-03-17 16:08:02.29551568 +0000 UTC m=+1873.708831727" observedRunningTime="2026-03-17 16:08:03.171109063 +0000 UTC m=+1874.584425120" watchObservedRunningTime="2026-03-17 16:08:03.22485223 +0000 UTC m=+1874.638168277" Mar 17 16:08:04 crc kubenswrapper[4767]: I0317 16:08:04.183967 4767 generic.go:334] "Generic (PLEG): container finished" podID="68a9bbf0-4d9d-4760-9aa7-3afc7aa20123" containerID="71cee4369dbc400c2680a693359d1227aa47dda82465b728603476ef258ef922" exitCode=0 Mar 17 16:08:04 crc kubenswrapper[4767]: I0317 16:08:04.184281 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562728-g6ss4" event={"ID":"68a9bbf0-4d9d-4760-9aa7-3afc7aa20123","Type":"ContainerDied","Data":"71cee4369dbc400c2680a693359d1227aa47dda82465b728603476ef258ef922"} Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.213325 4767 generic.go:334] "Generic (PLEG): container finished" podID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerID="d4a1f434e7c4a4ce983ca40097c549cebd416bd03f084596b19e3c868a329f18" exitCode=0 Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.213523 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfee8a92-09c3-424e-bdb9-973c50a10322","Type":"ContainerDied","Data":"d4a1f434e7c4a4ce983ca40097c549cebd416bd03f084596b19e3c868a329f18"} Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.450221 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.493460 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsdfz\" (UniqueName: \"kubernetes.io/projected/bfee8a92-09c3-424e-bdb9-973c50a10322-kube-api-access-bsdfz\") pod \"bfee8a92-09c3-424e-bdb9-973c50a10322\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.493901 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-scripts\") pod \"bfee8a92-09c3-424e-bdb9-973c50a10322\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.493985 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-config-data\") pod \"bfee8a92-09c3-424e-bdb9-973c50a10322\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.494331 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-run-httpd\") pod \"bfee8a92-09c3-424e-bdb9-973c50a10322\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.494362 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-log-httpd\") pod \"bfee8a92-09c3-424e-bdb9-973c50a10322\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.494383 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-combined-ca-bundle\") pod \"bfee8a92-09c3-424e-bdb9-973c50a10322\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.494510 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-sg-core-conf-yaml\") pod \"bfee8a92-09c3-424e-bdb9-973c50a10322\" (UID: \"bfee8a92-09c3-424e-bdb9-973c50a10322\") " Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.501965 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bfee8a92-09c3-424e-bdb9-973c50a10322" (UID: "bfee8a92-09c3-424e-bdb9-973c50a10322"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.502412 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bfee8a92-09c3-424e-bdb9-973c50a10322" (UID: "bfee8a92-09c3-424e-bdb9-973c50a10322"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.507644 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfee8a92-09c3-424e-bdb9-973c50a10322-kube-api-access-bsdfz" (OuterVolumeSpecName: "kube-api-access-bsdfz") pod "bfee8a92-09c3-424e-bdb9-973c50a10322" (UID: "bfee8a92-09c3-424e-bdb9-973c50a10322"). InnerVolumeSpecName "kube-api-access-bsdfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.508788 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-scripts" (OuterVolumeSpecName: "scripts") pod "bfee8a92-09c3-424e-bdb9-973c50a10322" (UID: "bfee8a92-09c3-424e-bdb9-973c50a10322"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.547713 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bfee8a92-09c3-424e-bdb9-973c50a10322" (UID: "bfee8a92-09c3-424e-bdb9-973c50a10322"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.599952 4767 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.599993 4767 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfee8a92-09c3-424e-bdb9-973c50a10322-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.600002 4767 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.600039 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsdfz\" (UniqueName: \"kubernetes.io/projected/bfee8a92-09c3-424e-bdb9-973c50a10322-kube-api-access-bsdfz\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.600048 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.678559 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfee8a92-09c3-424e-bdb9-973c50a10322" (UID: "bfee8a92-09c3-424e-bdb9-973c50a10322"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.688353 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562728-g6ss4" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.703631 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.736161 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-config-data" (OuterVolumeSpecName: "config-data") pod "bfee8a92-09c3-424e-bdb9-973c50a10322" (UID: "bfee8a92-09c3-424e-bdb9-973c50a10322"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.805864 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfzvr\" (UniqueName: \"kubernetes.io/projected/68a9bbf0-4d9d-4760-9aa7-3afc7aa20123-kube-api-access-vfzvr\") pod \"68a9bbf0-4d9d-4760-9aa7-3afc7aa20123\" (UID: \"68a9bbf0-4d9d-4760-9aa7-3afc7aa20123\") " Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.807545 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfee8a92-09c3-424e-bdb9-973c50a10322-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.811564 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68a9bbf0-4d9d-4760-9aa7-3afc7aa20123-kube-api-access-vfzvr" (OuterVolumeSpecName: "kube-api-access-vfzvr") pod "68a9bbf0-4d9d-4760-9aa7-3afc7aa20123" (UID: "68a9bbf0-4d9d-4760-9aa7-3afc7aa20123"). InnerVolumeSpecName "kube-api-access-vfzvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:05 crc kubenswrapper[4767]: I0317 16:08:05.911408 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfzvr\" (UniqueName: \"kubernetes.io/projected/68a9bbf0-4d9d-4760-9aa7-3afc7aa20123-kube-api-access-vfzvr\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.239367 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfee8a92-09c3-424e-bdb9-973c50a10322","Type":"ContainerDied","Data":"ca58a60067190c73fc0aa897136216d60546fbb2dc5d45adee71adac15fe35c2"} Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.239753 4767 scope.go:117] "RemoveContainer" containerID="ed4d3c6c450409bf7c4b5a8866365f5b5d9eaf77773bfe9a48bc9d2b4361696f" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.239939 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.249118 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562728-g6ss4" event={"ID":"68a9bbf0-4d9d-4760-9aa7-3afc7aa20123","Type":"ContainerDied","Data":"8c6a071f702fa0daf4e81323d7b5a7f2021b4daf1d9581d73200fa86e5e2aa78"} Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.249163 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c6a071f702fa0daf4e81323d7b5a7f2021b4daf1d9581d73200fa86e5e2aa78" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.249365 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562728-g6ss4" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.268092 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562722-prmjs"] Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.279822 4767 scope.go:117] "RemoveContainer" containerID="c2ba3abf835e8e32b378663c7dbe6060b1e565bed7bbcdb5aee104ef1acaec45" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.281434 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562722-prmjs"] Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.313187 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.338243 4767 scope.go:117] "RemoveContainer" containerID="d4a1f434e7c4a4ce983ca40097c549cebd416bd03f084596b19e3c868a329f18" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.338608 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.356415 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:06 crc kubenswrapper[4767]: E0317 16:08:06.357382 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="proxy-httpd" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.357406 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="proxy-httpd" Mar 17 16:08:06 crc kubenswrapper[4767]: E0317 16:08:06.357428 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="ceilometer-notification-agent" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.357437 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="ceilometer-notification-agent" Mar 17 16:08:06 crc kubenswrapper[4767]: E0317 16:08:06.357455 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="ceilometer-central-agent" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.357461 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="ceilometer-central-agent" Mar 17 16:08:06 crc kubenswrapper[4767]: E0317 16:08:06.357489 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a9bbf0-4d9d-4760-9aa7-3afc7aa20123" containerName="oc" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.357496 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a9bbf0-4d9d-4760-9aa7-3afc7aa20123" containerName="oc" Mar 17 16:08:06 crc kubenswrapper[4767]: E0317 16:08:06.357536 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="sg-core" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.357542 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="sg-core" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.357820 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="sg-core" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.357834 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="68a9bbf0-4d9d-4760-9aa7-3afc7aa20123" containerName="oc" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.357860 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="proxy-httpd" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.357874 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="ceilometer-notification-agent" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.357890 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" containerName="ceilometer-central-agent" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.361775 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.365912 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.369452 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.369614 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.395311 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.406112 4767 scope.go:117] "RemoveContainer" containerID="16ed2ddd83e90b96b95a234eaf96bcd80511e23d2a37782421f0d5b119b47c46" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.429037 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-config-data\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.429208 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-log-httpd\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.429303 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.429412 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.429458 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-scripts\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.429476 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.429566 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-run-httpd\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.429673 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x7cf\" (UniqueName: \"kubernetes.io/projected/bfd60c23-044e-4fae-8289-c460e98160fb-kube-api-access-4x7cf\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.534078 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-config-data\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.535696 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-log-httpd\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.535862 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.535977 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.536084 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-scripts\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.536113 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.536293 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-run-httpd\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.536371 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-log-httpd\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.536496 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x7cf\" (UniqueName: \"kubernetes.io/projected/bfd60c23-044e-4fae-8289-c460e98160fb-kube-api-access-4x7cf\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.536684 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-run-httpd\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.540717 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.540871 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.546140 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.546477 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-scripts\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.546975 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-config-data\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.568901 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x7cf\" (UniqueName: \"kubernetes.io/projected/bfd60c23-044e-4fae-8289-c460e98160fb-kube-api-access-4x7cf\") pod \"ceilometer-0\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " pod="openstack/ceilometer-0" Mar 17 16:08:06 crc kubenswrapper[4767]: I0317 16:08:06.693941 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.143361 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-657sp"] Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.165164 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-657sp"] Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.250012 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-4zghh"] Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.252404 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.301695 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-4zghh"] Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.328477 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.374083 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbxbp\" (UniqueName: \"kubernetes.io/projected/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-kube-api-access-kbxbp\") pod \"heat-db-sync-4zghh\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.374711 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-config-data\") pod \"heat-db-sync-4zghh\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.375780 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-combined-ca-bundle\") pod \"heat-db-sync-4zghh\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.375778 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e6a688d-7438-45a1-9aa7-b7042485163f" path="/var/lib/kubelet/pods/5e6a688d-7438-45a1-9aa7-b7042485163f/volumes" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.377044 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfee8a92-09c3-424e-bdb9-973c50a10322" path="/var/lib/kubelet/pods/bfee8a92-09c3-424e-bdb9-973c50a10322/volumes" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.378766 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd1f4000-0e5a-438c-80ad-2805f6f9d772" path="/var/lib/kubelet/pods/cd1f4000-0e5a-438c-80ad-2805f6f9d772/volumes" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.479992 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbxbp\" (UniqueName: \"kubernetes.io/projected/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-kube-api-access-kbxbp\") pod \"heat-db-sync-4zghh\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.480574 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-config-data\") pod \"heat-db-sync-4zghh\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.480659 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-combined-ca-bundle\") pod \"heat-db-sync-4zghh\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.489367 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-config-data\") pod \"heat-db-sync-4zghh\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.495346 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-combined-ca-bundle\") pod \"heat-db-sync-4zghh\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.501707 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbxbp\" (UniqueName: \"kubernetes.io/projected/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-kube-api-access-kbxbp\") pod \"heat-db-sync-4zghh\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:07 crc kubenswrapper[4767]: I0317 16:08:07.592272 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:08 crc kubenswrapper[4767]: W0317 16:08:08.130610 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd28d7ecb_5431_4abd_beb2_f4a4c22df3f3.slice/crio-fc90611c267553f377500baec18af3a5f7e6c01a09aed5bd5c8e43fb5ffc2ce9 WatchSource:0}: Error finding container fc90611c267553f377500baec18af3a5f7e6c01a09aed5bd5c8e43fb5ffc2ce9: Status 404 returned error can't find the container with id fc90611c267553f377500baec18af3a5f7e6c01a09aed5bd5c8e43fb5ffc2ce9 Mar 17 16:08:08 crc kubenswrapper[4767]: I0317 16:08:08.137626 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-4zghh"] Mar 17 16:08:08 crc kubenswrapper[4767]: I0317 16:08:08.302085 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfd60c23-044e-4fae-8289-c460e98160fb","Type":"ContainerStarted","Data":"c0c464a7fb59c6eb223bfdcc2dfb2d459a4d3790930cab1456c3fdb5472c211d"} Mar 17 16:08:08 crc kubenswrapper[4767]: I0317 16:08:08.302191 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfd60c23-044e-4fae-8289-c460e98160fb","Type":"ContainerStarted","Data":"48f21c04c829f8bdefdf5b3c9b408559f83a864496478a17cd58c73d9ce86f0b"} Mar 17 16:08:08 crc kubenswrapper[4767]: I0317 16:08:08.306081 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-4zghh" event={"ID":"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3","Type":"ContainerStarted","Data":"fc90611c267553f377500baec18af3a5f7e6c01a09aed5bd5c8e43fb5ffc2ce9"} Mar 17 16:08:08 crc kubenswrapper[4767]: I0317 16:08:08.968935 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 17 16:08:09 crc kubenswrapper[4767]: I0317 16:08:09.411322 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfd60c23-044e-4fae-8289-c460e98160fb","Type":"ContainerStarted","Data":"ea4836ecbfabd75ad9d519a69ca9c9fbb887d41dc80687e9b2b46a8672e4b753"} Mar 17 16:08:09 crc kubenswrapper[4767]: I0317 16:08:09.742060 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 17 16:08:09 crc kubenswrapper[4767]: I0317 16:08:09.802527 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lc99p" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="registry-server" probeResult="failure" output=< Mar 17 16:08:09 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:08:09 crc kubenswrapper[4767]: > Mar 17 16:08:10 crc kubenswrapper[4767]: I0317 16:08:10.386136 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfd60c23-044e-4fae-8289-c460e98160fb","Type":"ContainerStarted","Data":"646b42cbf2c3024b9786b9cbe2ad26b2dd6da5cfeabd1c21df005b490019d150"} Mar 17 16:08:10 crc kubenswrapper[4767]: I0317 16:08:10.610333 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 16:08:13 crc kubenswrapper[4767]: I0317 16:08:13.167319 4767 scope.go:117] "RemoveContainer" containerID="51f03ebbb40543f93b81461057db4f82070357de7e7ccc7088777523c2cc5c8b" Mar 17 16:08:13 crc kubenswrapper[4767]: I0317 16:08:13.260987 4767 scope.go:117] "RemoveContainer" containerID="e08a06a020465497aeb9d5f24945601e389e57714d515764ca2d594873a36436" Mar 17 16:08:13 crc kubenswrapper[4767]: I0317 16:08:13.707991 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:13 crc kubenswrapper[4767]: I0317 16:08:13.746803 4767 scope.go:117] "RemoveContainer" containerID="e23ded531abf1800f2b455e0669e8ae5b55542a41b015cf3583366ddff4093ec" Mar 17 16:08:14 crc kubenswrapper[4767]: I0317 16:08:14.355892 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:08:14 crc kubenswrapper[4767]: E0317 16:08:14.356801 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:08:14 crc kubenswrapper[4767]: I0317 16:08:14.548936 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfd60c23-044e-4fae-8289-c460e98160fb","Type":"ContainerStarted","Data":"d2ef1fd540f82d89b7ac4bddbfe0ab4115085ec3350e9e81116ef9f15ca6a057"} Mar 17 16:08:14 crc kubenswrapper[4767]: I0317 16:08:14.549243 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 16:08:14 crc kubenswrapper[4767]: I0317 16:08:14.557310 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-2" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerName="rabbitmq" containerID="cri-o://691fea73d6a12e81a25a85a1e6ada57dc114881e572a3370e586226a5dec638f" gracePeriod=604795 Mar 17 16:08:15 crc kubenswrapper[4767]: I0317 16:08:15.570349 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="proxy-httpd" containerID="cri-o://d2ef1fd540f82d89b7ac4bddbfe0ab4115085ec3350e9e81116ef9f15ca6a057" gracePeriod=30 Mar 17 16:08:15 crc kubenswrapper[4767]: I0317 16:08:15.570349 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="sg-core" containerID="cri-o://646b42cbf2c3024b9786b9cbe2ad26b2dd6da5cfeabd1c21df005b490019d150" gracePeriod=30 Mar 17 16:08:15 crc kubenswrapper[4767]: I0317 16:08:15.570346 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="ceilometer-central-agent" containerID="cri-o://c0c464a7fb59c6eb223bfdcc2dfb2d459a4d3790930cab1456c3fdb5472c211d" gracePeriod=30 Mar 17 16:08:15 crc kubenswrapper[4767]: I0317 16:08:15.570386 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="ceilometer-notification-agent" containerID="cri-o://ea4836ecbfabd75ad9d519a69ca9c9fbb887d41dc80687e9b2b46a8672e4b753" gracePeriod=30 Mar 17 16:08:16 crc kubenswrapper[4767]: I0317 16:08:16.663377 4767 generic.go:334] "Generic (PLEG): container finished" podID="bfd60c23-044e-4fae-8289-c460e98160fb" containerID="d2ef1fd540f82d89b7ac4bddbfe0ab4115085ec3350e9e81116ef9f15ca6a057" exitCode=0 Mar 17 16:08:16 crc kubenswrapper[4767]: I0317 16:08:16.663666 4767 generic.go:334] "Generic (PLEG): container finished" podID="bfd60c23-044e-4fae-8289-c460e98160fb" containerID="646b42cbf2c3024b9786b9cbe2ad26b2dd6da5cfeabd1c21df005b490019d150" exitCode=2 Mar 17 16:08:16 crc kubenswrapper[4767]: I0317 16:08:16.663693 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfd60c23-044e-4fae-8289-c460e98160fb","Type":"ContainerDied","Data":"d2ef1fd540f82d89b7ac4bddbfe0ab4115085ec3350e9e81116ef9f15ca6a057"} Mar 17 16:08:16 crc kubenswrapper[4767]: I0317 16:08:16.663725 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfd60c23-044e-4fae-8289-c460e98160fb","Type":"ContainerDied","Data":"646b42cbf2c3024b9786b9cbe2ad26b2dd6da5cfeabd1c21df005b490019d150"} Mar 17 16:08:17 crc kubenswrapper[4767]: I0317 16:08:17.052041 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" containerName="rabbitmq" containerID="cri-o://143c311e23cdc0a92bff7b528ae3f0e08fa41bb7556bfc2c67cccbe9c57c1306" gracePeriod=604794 Mar 17 16:08:17 crc kubenswrapper[4767]: I0317 16:08:17.685390 4767 generic.go:334] "Generic (PLEG): container finished" podID="bfd60c23-044e-4fae-8289-c460e98160fb" containerID="ea4836ecbfabd75ad9d519a69ca9c9fbb887d41dc80687e9b2b46a8672e4b753" exitCode=0 Mar 17 16:08:17 crc kubenswrapper[4767]: I0317 16:08:17.685567 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfd60c23-044e-4fae-8289-c460e98160fb","Type":"ContainerDied","Data":"ea4836ecbfabd75ad9d519a69ca9c9fbb887d41dc80687e9b2b46a8672e4b753"} Mar 17 16:08:19 crc kubenswrapper[4767]: I0317 16:08:19.791024 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lc99p" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="registry-server" probeResult="failure" output=< Mar 17 16:08:19 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:08:19 crc kubenswrapper[4767]: > Mar 17 16:08:19 crc kubenswrapper[4767]: I0317 16:08:19.829554 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.134:5671: connect: connection refused" Mar 17 16:08:20 crc kubenswrapper[4767]: I0317 16:08:20.325403 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.135:5671: connect: connection refused" Mar 17 16:08:20 crc kubenswrapper[4767]: I0317 16:08:20.791453 4767 generic.go:334] "Generic (PLEG): container finished" podID="bfd60c23-044e-4fae-8289-c460e98160fb" containerID="c0c464a7fb59c6eb223bfdcc2dfb2d459a4d3790930cab1456c3fdb5472c211d" exitCode=0 Mar 17 16:08:20 crc kubenswrapper[4767]: I0317 16:08:20.791712 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfd60c23-044e-4fae-8289-c460e98160fb","Type":"ContainerDied","Data":"c0c464a7fb59c6eb223bfdcc2dfb2d459a4d3790930cab1456c3fdb5472c211d"} Mar 17 16:08:21 crc kubenswrapper[4767]: I0317 16:08:21.824913 4767 generic.go:334] "Generic (PLEG): container finished" podID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerID="691fea73d6a12e81a25a85a1e6ada57dc114881e572a3370e586226a5dec638f" exitCode=0 Mar 17 16:08:21 crc kubenswrapper[4767]: I0317 16:08:21.824997 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a082b2bf-8b94-40e7-be0c-be64f75a4c3e","Type":"ContainerDied","Data":"691fea73d6a12e81a25a85a1e6ada57dc114881e572a3370e586226a5dec638f"} Mar 17 16:08:23 crc kubenswrapper[4767]: I0317 16:08:23.879751 4767 generic.go:334] "Generic (PLEG): container finished" podID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" containerID="143c311e23cdc0a92bff7b528ae3f0e08fa41bb7556bfc2c67cccbe9c57c1306" exitCode=0 Mar 17 16:08:23 crc kubenswrapper[4767]: I0317 16:08:23.879972 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e","Type":"ContainerDied","Data":"143c311e23cdc0a92bff7b528ae3f0e08fa41bb7556bfc2c67cccbe9c57c1306"} Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.254915 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=14.878962175 podStartE2EDuration="20.254888014s" podCreationTimestamp="2026-03-17 16:08:06 +0000 UTC" firstStartedPulling="2026-03-17 16:08:07.311663159 +0000 UTC m=+1878.724979206" lastFinishedPulling="2026-03-17 16:08:12.687588998 +0000 UTC m=+1884.100905045" observedRunningTime="2026-03-17 16:08:14.593655466 +0000 UTC m=+1886.006971533" watchObservedRunningTime="2026-03-17 16:08:26.254888014 +0000 UTC m=+1897.668204061" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.266145 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-hm5mf"] Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.269203 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.278753 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.314309 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.314743 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pztkq\" (UniqueName: \"kubernetes.io/projected/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-kube-api-access-pztkq\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.314811 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.314868 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.315477 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.315604 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-config\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.315712 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.328734 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-hm5mf"] Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.355523 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:08:26 crc kubenswrapper[4767]: E0317 16:08:26.356308 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.418437 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.418521 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pztkq\" (UniqueName: \"kubernetes.io/projected/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-kube-api-access-pztkq\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.418623 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.418712 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.419027 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.419104 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-config\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.419206 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.419634 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.421077 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.421997 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-config\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.422381 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.422483 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.423262 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.497351 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pztkq\" (UniqueName: \"kubernetes.io/projected/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-kube-api-access-pztkq\") pod \"dnsmasq-dns-7d84b4d45c-hm5mf\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:26 crc kubenswrapper[4767]: I0317 16:08:26.619642 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:29 crc kubenswrapper[4767]: I0317 16:08:29.799793 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lc99p" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="registry-server" probeResult="failure" output=< Mar 17 16:08:29 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:08:29 crc kubenswrapper[4767]: > Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.584378 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.667791 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-erlang-cookie-secret\") pod \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.667867 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-confd\") pod \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.667930 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-tls\") pod \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.668324 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-plugins-conf\") pod \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.668472 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-erlang-cookie\") pod \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.668510 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-config-data\") pod \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.668568 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrjld\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-kube-api-access-lrjld\") pod \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.668664 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-pod-info\") pod \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.668706 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-plugins\") pod \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.677336 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a082b2bf-8b94-40e7-be0c-be64f75a4c3e" (UID: "a082b2bf-8b94-40e7-be0c-be64f75a4c3e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.678642 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a082b2bf-8b94-40e7-be0c-be64f75a4c3e" (UID: "a082b2bf-8b94-40e7-be0c-be64f75a4c3e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.679085 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a082b2bf-8b94-40e7-be0c-be64f75a4c3e" (UID: "a082b2bf-8b94-40e7-be0c-be64f75a4c3e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.679591 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a082b2bf-8b94-40e7-be0c-be64f75a4c3e" (UID: "a082b2bf-8b94-40e7-be0c-be64f75a4c3e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.694213 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a082b2bf-8b94-40e7-be0c-be64f75a4c3e" (UID: "a082b2bf-8b94-40e7-be0c-be64f75a4c3e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.695226 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-kube-api-access-lrjld" (OuterVolumeSpecName: "kube-api-access-lrjld") pod "a082b2bf-8b94-40e7-be0c-be64f75a4c3e" (UID: "a082b2bf-8b94-40e7-be0c-be64f75a4c3e"). InnerVolumeSpecName "kube-api-access-lrjld". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.704553 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-pod-info" (OuterVolumeSpecName: "pod-info") pod "a082b2bf-8b94-40e7-be0c-be64f75a4c3e" (UID: "a082b2bf-8b94-40e7-be0c-be64f75a4c3e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.750959 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-config-data" (OuterVolumeSpecName: "config-data") pod "a082b2bf-8b94-40e7-be0c-be64f75a4c3e" (UID: "a082b2bf-8b94-40e7-be0c-be64f75a4c3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.774114 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-server-conf\") pod \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.775178 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f\") pod \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\" (UID: \"a082b2bf-8b94-40e7-be0c-be64f75a4c3e\") " Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.775719 4767 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.775748 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.775757 4767 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.775766 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.775778 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.775786 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrjld\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-kube-api-access-lrjld\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.775794 4767 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-pod-info\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:30 crc kubenswrapper[4767]: I0317 16:08:30.775802 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.241277 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f" (OuterVolumeSpecName: "persistence") pod "a082b2bf-8b94-40e7-be0c-be64f75a4c3e" (UID: "a082b2bf-8b94-40e7-be0c-be64f75a4c3e"). InnerVolumeSpecName "pvc-6bac37c2-c76e-45e2-82cf-78846509636f". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.272917 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-server-conf" (OuterVolumeSpecName: "server-conf") pod "a082b2bf-8b94-40e7-be0c-be64f75a4c3e" (UID: "a082b2bf-8b94-40e7-be0c-be64f75a4c3e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.276504 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a082b2bf-8b94-40e7-be0c-be64f75a4c3e","Type":"ContainerDied","Data":"e8c41589e69f8552b3017809e34206b546ad0edfaaa82bfd1e7e2322f1b3691f"} Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.276566 4767 scope.go:117] "RemoveContainer" containerID="691fea73d6a12e81a25a85a1e6ada57dc114881e572a3370e586226a5dec638f" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.276720 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.314872 4767 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-6bac37c2-c76e-45e2-82cf-78846509636f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f\") on node \"crc\" " Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.314945 4767 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-server-conf\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.352583 4767 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.352876 4767 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-6bac37c2-c76e-45e2-82cf-78846509636f" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f") on node "crc" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.357264 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a082b2bf-8b94-40e7-be0c-be64f75a4c3e" (UID: "a082b2bf-8b94-40e7-be0c-be64f75a4c3e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.419742 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a082b2bf-8b94-40e7-be0c-be64f75a4c3e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.419813 4767 reconciler_common.go:293] "Volume detached for volume \"pvc-6bac37c2-c76e-45e2-82cf-78846509636f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.641256 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.671179 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.728473 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-2"] Mar 17 16:08:31 crc kubenswrapper[4767]: E0317 16:08:31.729397 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerName="rabbitmq" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.729426 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerName="rabbitmq" Mar 17 16:08:31 crc kubenswrapper[4767]: E0317 16:08:31.729459 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerName="setup-container" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.729469 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerName="setup-container" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.729817 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerName="rabbitmq" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.731783 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.770974 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.935929 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a298936e-5022-44de-9d7a-65f67ec76087-pod-info\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.936383 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a298936e-5022-44de-9d7a-65f67ec76087-config-data\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.936427 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a298936e-5022-44de-9d7a-65f67ec76087-server-conf\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.936474 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blcqm\" (UniqueName: \"kubernetes.io/projected/a298936e-5022-44de-9d7a-65f67ec76087-kube-api-access-blcqm\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.936762 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.936986 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.937044 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.937370 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a298936e-5022-44de-9d7a-65f67ec76087-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.937451 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.937580 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6bac37c2-c76e-45e2-82cf-78846509636f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:31 crc kubenswrapper[4767]: I0317 16:08:31.937701 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a298936e-5022-44de-9d7a-65f67ec76087-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.041446 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a298936e-5022-44de-9d7a-65f67ec76087-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.041968 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.042208 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6bac37c2-c76e-45e2-82cf-78846509636f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.042345 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a298936e-5022-44de-9d7a-65f67ec76087-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.042591 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a298936e-5022-44de-9d7a-65f67ec76087-pod-info\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.042763 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a298936e-5022-44de-9d7a-65f67ec76087-config-data\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.042877 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a298936e-5022-44de-9d7a-65f67ec76087-server-conf\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.042997 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blcqm\" (UniqueName: \"kubernetes.io/projected/a298936e-5022-44de-9d7a-65f67ec76087-kube-api-access-blcqm\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.043165 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.043367 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.043559 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.044481 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a298936e-5022-44de-9d7a-65f67ec76087-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.044494 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.044879 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a298936e-5022-44de-9d7a-65f67ec76087-server-conf\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.045551 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.045628 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a298936e-5022-44de-9d7a-65f67ec76087-config-data\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.051381 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.051438 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6bac37c2-c76e-45e2-82cf-78846509636f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/86d7f32ead6e09779828e242e22d0a054b17ff2cb152b126a42966414d82f3c6/globalmount\"" pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.053608 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.059803 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a298936e-5022-44de-9d7a-65f67ec76087-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.070283 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a298936e-5022-44de-9d7a-65f67ec76087-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.071868 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a298936e-5022-44de-9d7a-65f67ec76087-pod-info\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.081988 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blcqm\" (UniqueName: \"kubernetes.io/projected/a298936e-5022-44de-9d7a-65f67ec76087-kube-api-access-blcqm\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.138343 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6bac37c2-c76e-45e2-82cf-78846509636f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bac37c2-c76e-45e2-82cf-78846509636f\") pod \"rabbitmq-server-2\" (UID: \"a298936e-5022-44de-9d7a-65f67ec76087\") " pod="openstack/rabbitmq-server-2" Mar 17 16:08:32 crc kubenswrapper[4767]: I0317 16:08:32.359959 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 17 16:08:33 crc kubenswrapper[4767]: I0317 16:08:33.387157 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" path="/var/lib/kubelet/pods/a082b2bf-8b94-40e7-be0c-be64f75a4c3e/volumes" Mar 17 16:08:34 crc kubenswrapper[4767]: I0317 16:08:34.827457 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="a082b2bf-8b94-40e7-be0c-be64f75a4c3e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.134:5671: i/o timeout" Mar 17 16:08:35 crc kubenswrapper[4767]: I0317 16:08:35.325708 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.135:5671: i/o timeout" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.340365 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.354762 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.412090 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.415903 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.421987 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x7cf\" (UniqueName: \"kubernetes.io/projected/bfd60c23-044e-4fae-8289-c460e98160fb-kube-api-access-4x7cf\") pod \"bfd60c23-044e-4fae-8289-c460e98160fb\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.422048 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-config-data\") pod \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.422095 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-ceilometer-tls-certs\") pod \"bfd60c23-044e-4fae-8289-c460e98160fb\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.429842 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\") pod \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430029 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-plugins-conf\") pod \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430094 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-erlang-cookie-secret\") pod \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430121 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-run-httpd\") pod \"bfd60c23-044e-4fae-8289-c460e98160fb\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430186 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-server-conf\") pod \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430244 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-erlang-cookie\") pod \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430313 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-combined-ca-bundle\") pod \"bfd60c23-044e-4fae-8289-c460e98160fb\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430347 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-config-data\") pod \"bfd60c23-044e-4fae-8289-c460e98160fb\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430415 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-pod-info\") pod \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430430 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-plugins\") pod \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430446 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-scripts\") pod \"bfd60c23-044e-4fae-8289-c460e98160fb\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430465 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nwj5\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-kube-api-access-8nwj5\") pod \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430495 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-log-httpd\") pod \"bfd60c23-044e-4fae-8289-c460e98160fb\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430529 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-tls\") pod \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430616 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-confd\") pod \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\" (UID: \"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.430648 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-sg-core-conf-yaml\") pod \"bfd60c23-044e-4fae-8289-c460e98160fb\" (UID: \"bfd60c23-044e-4fae-8289-c460e98160fb\") " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.432501 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bfd60c23-044e-4fae-8289-c460e98160fb" (UID: "bfd60c23-044e-4fae-8289-c460e98160fb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.432642 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bfd60c23-044e-4fae-8289-c460e98160fb" (UID: "bfd60c23-044e-4fae-8289-c460e98160fb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.433451 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" (UID: "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.443345 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfd60c23-044e-4fae-8289-c460e98160fb","Type":"ContainerDied","Data":"48f21c04c829f8bdefdf5b3c9b408559f83a864496478a17cd58c73d9ce86f0b"} Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.443400 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef2226f0-7710-4a82-84ab-b4ce37f8bf2e","Type":"ContainerDied","Data":"3f40325b824810ca3a4f2de9038858fc4da068651267cf17d245c9681d1b797c"} Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.460118 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" (UID: "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.460379 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" (UID: "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.468707 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" (UID: "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.473582 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-pod-info" (OuterVolumeSpecName: "pod-info") pod "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" (UID: "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.473828 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfd60c23-044e-4fae-8289-c460e98160fb-kube-api-access-4x7cf" (OuterVolumeSpecName: "kube-api-access-4x7cf") pod "bfd60c23-044e-4fae-8289-c460e98160fb" (UID: "bfd60c23-044e-4fae-8289-c460e98160fb"). InnerVolumeSpecName "kube-api-access-4x7cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.488915 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-scripts" (OuterVolumeSpecName: "scripts") pod "bfd60c23-044e-4fae-8289-c460e98160fb" (UID: "bfd60c23-044e-4fae-8289-c460e98160fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.504571 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-kube-api-access-8nwj5" (OuterVolumeSpecName: "kube-api-access-8nwj5") pod "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" (UID: "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e"). InnerVolumeSpecName "kube-api-access-8nwj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.535295 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x7cf\" (UniqueName: \"kubernetes.io/projected/bfd60c23-044e-4fae-8289-c460e98160fb-kube-api-access-4x7cf\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.535332 4767 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.535343 4767 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.535352 4767 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.535361 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.535370 4767 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-pod-info\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.535377 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.535385 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.535399 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nwj5\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-kube-api-access-8nwj5\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.535407 4767 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfd60c23-044e-4fae-8289-c460e98160fb-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.548665 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" (UID: "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.553607 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180" (OuterVolumeSpecName: "persistence") pod "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" (UID: "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e"). InnerVolumeSpecName "pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.563440 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-config-data" (OuterVolumeSpecName: "config-data") pod "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" (UID: "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.640353 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.640756 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.640993 4767 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\") on node \"crc\" " Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.644640 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bfd60c23-044e-4fae-8289-c460e98160fb" (UID: "bfd60c23-044e-4fae-8289-c460e98160fb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.671767 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "bfd60c23-044e-4fae-8289-c460e98160fb" (UID: "bfd60c23-044e-4fae-8289-c460e98160fb"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.687785 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-server-conf" (OuterVolumeSpecName: "server-conf") pod "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" (UID: "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.724464 4767 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.724971 4767 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180") on node "crc" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.727018 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfd60c23-044e-4fae-8289-c460e98160fb" (UID: "bfd60c23-044e-4fae-8289-c460e98160fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.750610 4767 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-server-conf\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.750666 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.750679 4767 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.750691 4767 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.750707 4767 reconciler_common.go:293] "Volume detached for volume \"pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.772350 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" (UID: "ef2226f0-7710-4a82-84ab-b4ce37f8bf2e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.791247 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-config-data" (OuterVolumeSpecName: "config-data") pod "bfd60c23-044e-4fae-8289-c460e98160fb" (UID: "bfd60c23-044e-4fae-8289-c460e98160fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.867922 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfd60c23-044e-4fae-8289-c460e98160fb-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:37 crc kubenswrapper[4767]: I0317 16:08:37.868318 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.060120 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.077739 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.097267 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.113266 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 16:08:38 crc kubenswrapper[4767]: E0317 16:08:38.127230 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Mar 17 16:08:38 crc kubenswrapper[4767]: E0317 16:08:38.127318 4767 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Mar 17 16:08:38 crc kubenswrapper[4767]: E0317 16:08:38.127468 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kbxbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-4zghh_openstack(d28d7ecb-5431-4abd-beb2-f4a4c22df3f3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:08:38 crc kubenswrapper[4767]: E0317 16:08:38.129288 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-4zghh" podUID="d28d7ecb-5431-4abd-beb2-f4a4c22df3f3" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.131265 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:38 crc kubenswrapper[4767]: E0317 16:08:38.132035 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" containerName="setup-container" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.132055 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" containerName="setup-container" Mar 17 16:08:38 crc kubenswrapper[4767]: E0317 16:08:38.132067 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="ceilometer-notification-agent" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.132073 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="ceilometer-notification-agent" Mar 17 16:08:38 crc kubenswrapper[4767]: E0317 16:08:38.132094 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="sg-core" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.132100 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="sg-core" Mar 17 16:08:38 crc kubenswrapper[4767]: E0317 16:08:38.132112 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="proxy-httpd" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.132118 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="proxy-httpd" Mar 17 16:08:38 crc kubenswrapper[4767]: E0317 16:08:38.132130 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="ceilometer-central-agent" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.132137 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="ceilometer-central-agent" Mar 17 16:08:38 crc kubenswrapper[4767]: E0317 16:08:38.132157 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" containerName="rabbitmq" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.132163 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" containerName="rabbitmq" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.132492 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="ceilometer-central-agent" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.132514 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="ceilometer-notification-agent" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.132522 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="sg-core" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.132537 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="proxy-httpd" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.132544 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" containerName="rabbitmq" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.135090 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.138346 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.143846 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.143918 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.147154 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.156022 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.156734 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.156929 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.157092 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.157432 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.157612 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.157842 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.158126 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-cczbq" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.158430 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.180078 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cce1c59c-7c85-4922-bbbc-462261019bc9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.180227 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cce1c59c-7c85-4922-bbbc-462261019bc9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.180256 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.180282 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.180342 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cce1c59c-7c85-4922-bbbc-462261019bc9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.180373 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nhs5\" (UniqueName: \"kubernetes.io/projected/cce1c59c-7c85-4922-bbbc-462261019bc9-kube-api-access-6nhs5\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.180412 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.180430 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cce1c59c-7c85-4922-bbbc-462261019bc9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.180455 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cce1c59c-7c85-4922-bbbc-462261019bc9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.180506 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.180537 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.181386 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.199563 4767 scope.go:117] "RemoveContainer" containerID="a150a9a169d4d790c0367da0e3743c173a7add3e0702539d16e0be602f1d139a" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.283938 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cce1c59c-7c85-4922-bbbc-462261019bc9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284034 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284108 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nhs5\" (UniqueName: \"kubernetes.io/projected/cce1c59c-7c85-4922-bbbc-462261019bc9-kube-api-access-6nhs5\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284258 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284310 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284326 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cce1c59c-7c85-4922-bbbc-462261019bc9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284357 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cce1c59c-7c85-4922-bbbc-462261019bc9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284426 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d24c721d-4968-477f-ba8e-23f6d31629d2-run-httpd\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284452 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284492 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284564 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cce1c59c-7c85-4922-bbbc-462261019bc9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284586 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-scripts\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284627 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284676 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d24c721d-4968-477f-ba8e-23f6d31629d2-log-httpd\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284713 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-config-data\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284749 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cce1c59c-7c85-4922-bbbc-462261019bc9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284768 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284789 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.284809 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4grn7\" (UniqueName: \"kubernetes.io/projected/d24c721d-4968-477f-ba8e-23f6d31629d2-kube-api-access-4grn7\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.285683 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.285986 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.287807 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cce1c59c-7c85-4922-bbbc-462261019bc9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.288391 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cce1c59c-7c85-4922-bbbc-462261019bc9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.289710 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cce1c59c-7c85-4922-bbbc-462261019bc9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.293930 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cce1c59c-7c85-4922-bbbc-462261019bc9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.295907 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.296401 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.296651 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ceeeb535379e2ca1a9b6c52455f69317dc04eb485e1665f5be586667ac13ba46/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.301847 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cce1c59c-7c85-4922-bbbc-462261019bc9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.310972 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cce1c59c-7c85-4922-bbbc-462261019bc9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.315969 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nhs5\" (UniqueName: \"kubernetes.io/projected/cce1c59c-7c85-4922-bbbc-462261019bc9-kube-api-access-6nhs5\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.387101 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d24c721d-4968-477f-ba8e-23f6d31629d2-run-httpd\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.388559 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-scripts\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.388710 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.388842 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d24c721d-4968-477f-ba8e-23f6d31629d2-log-httpd\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.388949 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-config-data\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.389061 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4grn7\" (UniqueName: \"kubernetes.io/projected/d24c721d-4968-477f-ba8e-23f6d31629d2-kube-api-access-4grn7\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.388132 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d24c721d-4968-477f-ba8e-23f6d31629d2-run-httpd\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.387670 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ef4a5ce9-a0e0-4d52-965e-cad992a6d180\") pod \"rabbitmq-cell1-server-0\" (UID: \"cce1c59c-7c85-4922-bbbc-462261019bc9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.389335 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.389477 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.390052 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d24c721d-4968-477f-ba8e-23f6d31629d2-log-httpd\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.394768 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.394988 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-config-data\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.395718 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.396082 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.399277 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d24c721d-4968-477f-ba8e-23f6d31629d2-scripts\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.419800 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4grn7\" (UniqueName: \"kubernetes.io/projected/d24c721d-4968-477f-ba8e-23f6d31629d2-kube-api-access-4grn7\") pod \"ceilometer-0\" (UID: \"d24c721d-4968-477f-ba8e-23f6d31629d2\") " pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: E0317 16:08:38.484454 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-4zghh" podUID="d28d7ecb-5431-4abd-beb2-f4a4c22df3f3" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.488482 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.539696 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.561404 4767 scope.go:117] "RemoveContainer" containerID="d2ef1fd540f82d89b7ac4bddbfe0ab4115085ec3350e9e81116ef9f15ca6a057" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.646308 4767 scope.go:117] "RemoveContainer" containerID="646b42cbf2c3024b9786b9cbe2ad26b2dd6da5cfeabd1c21df005b490019d150" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.726575 4767 scope.go:117] "RemoveContainer" containerID="ea4836ecbfabd75ad9d519a69ca9c9fbb887d41dc80687e9b2b46a8672e4b753" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.773559 4767 scope.go:117] "RemoveContainer" containerID="c0c464a7fb59c6eb223bfdcc2dfb2d459a4d3790930cab1456c3fdb5472c211d" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.863987 4767 scope.go:117] "RemoveContainer" containerID="143c311e23cdc0a92bff7b528ae3f0e08fa41bb7556bfc2c67cccbe9c57c1306" Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.922459 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-hm5mf"] Mar 17 16:08:38 crc kubenswrapper[4767]: I0317 16:08:38.929853 4767 scope.go:117] "RemoveContainer" containerID="db91577d9d962897e9869c843454246c300eb4ac87d6b82c10120049bf6d2916" Mar 17 16:08:38 crc kubenswrapper[4767]: W0317 16:08:38.960230 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda40b3eb6_24e8_4aa2_bbbb_ae1233efcbba.slice/crio-72e6558b1fd18e7c0bc1b407e59139269635f6f0ce483cd8296093c17fc7ca66 WatchSource:0}: Error finding container 72e6558b1fd18e7c0bc1b407e59139269635f6f0ce483cd8296093c17fc7ca66: Status 404 returned error can't find the container with id 72e6558b1fd18e7c0bc1b407e59139269635f6f0ce483cd8296093c17fc7ca66 Mar 17 16:08:39 crc kubenswrapper[4767]: I0317 16:08:39.021604 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 17 16:08:39 crc kubenswrapper[4767]: W0317 16:08:39.022959 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda298936e_5022_44de_9d7a_65f67ec76087.slice/crio-7fa94c93bf367a45c064cdf740a2b0019fc68c7d30277a25c1aaa34fa8a7d70a WatchSource:0}: Error finding container 7fa94c93bf367a45c064cdf740a2b0019fc68c7d30277a25c1aaa34fa8a7d70a: Status 404 returned error can't find the container with id 7fa94c93bf367a45c064cdf740a2b0019fc68c7d30277a25c1aaa34fa8a7d70a Mar 17 16:08:39 crc kubenswrapper[4767]: I0317 16:08:39.296920 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 16:08:39 crc kubenswrapper[4767]: I0317 16:08:39.314875 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 16:08:39 crc kubenswrapper[4767]: I0317 16:08:39.624633 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" path="/var/lib/kubelet/pods/bfd60c23-044e-4fae-8289-c460e98160fb/volumes" Mar 17 16:08:39 crc kubenswrapper[4767]: I0317 16:08:39.635867 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef2226f0-7710-4a82-84ab-b4ce37f8bf2e" path="/var/lib/kubelet/pods/ef2226f0-7710-4a82-84ab-b4ce37f8bf2e/volumes" Mar 17 16:08:39 crc kubenswrapper[4767]: I0317 16:08:39.637952 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a298936e-5022-44de-9d7a-65f67ec76087","Type":"ContainerStarted","Data":"7fa94c93bf367a45c064cdf740a2b0019fc68c7d30277a25c1aaa34fa8a7d70a"} Mar 17 16:08:39 crc kubenswrapper[4767]: I0317 16:08:39.638135 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" event={"ID":"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba","Type":"ContainerStarted","Data":"72e6558b1fd18e7c0bc1b407e59139269635f6f0ce483cd8296093c17fc7ca66"} Mar 17 16:08:39 crc kubenswrapper[4767]: I0317 16:08:39.652191 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cce1c59c-7c85-4922-bbbc-462261019bc9","Type":"ContainerStarted","Data":"abf97e2bc006248fa60f85c0b79d7904854a94e9f61aa4924129346e186dad64"} Mar 17 16:08:39 crc kubenswrapper[4767]: I0317 16:08:39.686037 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d24c721d-4968-477f-ba8e-23f6d31629d2","Type":"ContainerStarted","Data":"d925825845bcbc1d54dd173ed67b9559ca136bd690cd9b83f9875b3fc6765e59"} Mar 17 16:08:39 crc kubenswrapper[4767]: I0317 16:08:39.812046 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lc99p" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="registry-server" probeResult="failure" output=< Mar 17 16:08:39 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:08:39 crc kubenswrapper[4767]: > Mar 17 16:08:40 crc kubenswrapper[4767]: I0317 16:08:40.731308 4767 generic.go:334] "Generic (PLEG): container finished" podID="a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" containerID="e96194c126e9320fbc653d93eb1880202233241335ea78ae4ab4b7862abce819" exitCode=0 Mar 17 16:08:40 crc kubenswrapper[4767]: I0317 16:08:40.731683 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" event={"ID":"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba","Type":"ContainerDied","Data":"e96194c126e9320fbc653d93eb1880202233241335ea78ae4ab4b7862abce819"} Mar 17 16:08:41 crc kubenswrapper[4767]: I0317 16:08:41.355485 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:08:41 crc kubenswrapper[4767]: E0317 16:08:41.356266 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:08:42 crc kubenswrapper[4767]: I0317 16:08:42.761942 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a298936e-5022-44de-9d7a-65f67ec76087","Type":"ContainerStarted","Data":"4338209039d729077a65c9693399c1133c96fd0b7fa40d3a483955d6e2e4f031"} Mar 17 16:08:42 crc kubenswrapper[4767]: I0317 16:08:42.765727 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" event={"ID":"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba","Type":"ContainerStarted","Data":"95da0a180a4468719b26f8b482508ee25c55c837077e708b354b1bcc56c9edb7"} Mar 17 16:08:42 crc kubenswrapper[4767]: I0317 16:08:42.765913 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:42 crc kubenswrapper[4767]: I0317 16:08:42.767902 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cce1c59c-7c85-4922-bbbc-462261019bc9","Type":"ContainerStarted","Data":"7d505a0d7bbd0bd3e6942d51090cd0e1bfebadb499cf0b186996aa2a0c474706"} Mar 17 16:08:42 crc kubenswrapper[4767]: I0317 16:08:42.862895 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" podStartSLOduration=16.862864329 podStartE2EDuration="16.862864329s" podCreationTimestamp="2026-03-17 16:08:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:08:42.860584854 +0000 UTC m=+1914.273900921" watchObservedRunningTime="2026-03-17 16:08:42.862864329 +0000 UTC m=+1914.276180376" Mar 17 16:08:44 crc kubenswrapper[4767]: I0317 16:08:44.821687 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d24c721d-4968-477f-ba8e-23f6d31629d2","Type":"ContainerStarted","Data":"bb348c2b124362aa1be574374c3cd321b25ea76e59aee8aba55f432d9c45538f"} Mar 17 16:08:45 crc kubenswrapper[4767]: I0317 16:08:45.853057 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d24c721d-4968-477f-ba8e-23f6d31629d2","Type":"ContainerStarted","Data":"5562dce513514a2267dcfcae3d07e48a4a2caad9139f513dee081e982e81032d"} Mar 17 16:08:46 crc kubenswrapper[4767]: I0317 16:08:46.622517 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:46 crc kubenswrapper[4767]: I0317 16:08:46.709888 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt"] Mar 17 16:08:46 crc kubenswrapper[4767]: I0317 16:08:46.710245 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" podUID="a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" containerName="dnsmasq-dns" containerID="cri-o://dd041368d5d907e35d0a9627903cda8654a62f7b46643599a6bcd662cfa25d67" gracePeriod=10 Mar 17 16:08:46 crc kubenswrapper[4767]: I0317 16:08:46.878105 4767 generic.go:334] "Generic (PLEG): container finished" podID="a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" containerID="dd041368d5d907e35d0a9627903cda8654a62f7b46643599a6bcd662cfa25d67" exitCode=0 Mar 17 16:08:46 crc kubenswrapper[4767]: I0317 16:08:46.878518 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" event={"ID":"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52","Type":"ContainerDied","Data":"dd041368d5d907e35d0a9627903cda8654a62f7b46643599a6bcd662cfa25d67"} Mar 17 16:08:46 crc kubenswrapper[4767]: I0317 16:08:46.886718 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d24c721d-4968-477f-ba8e-23f6d31629d2","Type":"ContainerStarted","Data":"e6afa628e312a9c36d79f16de8cddb90a5e9dd78ef1a6332f19b840e9ae4e901"} Mar 17 16:08:46 crc kubenswrapper[4767]: I0317 16:08:46.960891 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-jxg2g"] Mar 17 16:08:46 crc kubenswrapper[4767]: I0317 16:08:46.963387 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:46 crc kubenswrapper[4767]: I0317 16:08:46.988966 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-jxg2g"] Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.018091 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.018503 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psccn\" (UniqueName: \"kubernetes.io/projected/04dde071-2bb2-4e33-8253-d80c863b55e9-kube-api-access-psccn\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.018800 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.018949 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.019127 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-config\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.019267 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.019667 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.035967 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" podUID="a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.9:5353: connect: connection refused" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.124137 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.124239 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.124313 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-config\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.124332 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.124434 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.124568 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.124589 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psccn\" (UniqueName: \"kubernetes.io/projected/04dde071-2bb2-4e33-8253-d80c863b55e9-kube-api-access-psccn\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.125564 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.125662 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.125791 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.126054 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-config\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.126144 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.126512 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04dde071-2bb2-4e33-8253-d80c863b55e9-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.161953 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psccn\" (UniqueName: \"kubernetes.io/projected/04dde071-2bb2-4e33-8253-d80c863b55e9-kube-api-access-psccn\") pod \"dnsmasq-dns-6f6df4f56c-jxg2g\" (UID: \"04dde071-2bb2-4e33-8253-d80c863b55e9\") " pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.283673 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.482009 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.545762 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-svc\") pod \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.546047 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-sb\") pod \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.546309 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64z9w\" (UniqueName: \"kubernetes.io/projected/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-kube-api-access-64z9w\") pod \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.546427 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-config\") pod \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.546555 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-swift-storage-0\") pod \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.546747 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-nb\") pod \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\" (UID: \"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52\") " Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.559312 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-kube-api-access-64z9w" (OuterVolumeSpecName: "kube-api-access-64z9w") pod "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" (UID: "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52"). InnerVolumeSpecName "kube-api-access-64z9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.650349 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" (UID: "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.650959 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64z9w\" (UniqueName: \"kubernetes.io/projected/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-kube-api-access-64z9w\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.651010 4767 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.658591 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-config" (OuterVolumeSpecName: "config") pod "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" (UID: "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.666809 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" (UID: "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.679726 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" (UID: "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.697863 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" (UID: "a87e499b-99c4-440b-8f1f-2d9e6ef1cd52"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.754479 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.754777 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.754879 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.755017 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.907225 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" event={"ID":"a87e499b-99c4-440b-8f1f-2d9e6ef1cd52","Type":"ContainerDied","Data":"72b84b762715964cba4e0b080485b73617ff7e0a0010833944f58f527307b8f1"} Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.908033 4767 scope.go:117] "RemoveContainer" containerID="dd041368d5d907e35d0a9627903cda8654a62f7b46643599a6bcd662cfa25d67" Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.908574 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt" Mar 17 16:08:47 crc kubenswrapper[4767]: W0317 16:08:47.916027 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04dde071_2bb2_4e33_8253_d80c863b55e9.slice/crio-433a4e8d3aec0543bf9ce3f1b20bed5723a1a4ccc14e08779e02570f225e5874 WatchSource:0}: Error finding container 433a4e8d3aec0543bf9ce3f1b20bed5723a1a4ccc14e08779e02570f225e5874: Status 404 returned error can't find the container with id 433a4e8d3aec0543bf9ce3f1b20bed5723a1a4ccc14e08779e02570f225e5874 Mar 17 16:08:47 crc kubenswrapper[4767]: I0317 16:08:47.935944 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-jxg2g"] Mar 17 16:08:48 crc kubenswrapper[4767]: I0317 16:08:48.195085 4767 scope.go:117] "RemoveContainer" containerID="0769b432f754f478e1498527b6409bf1a54a6686b7d987b80f89e082fecce414" Mar 17 16:08:48 crc kubenswrapper[4767]: I0317 16:08:48.239535 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt"] Mar 17 16:08:48 crc kubenswrapper[4767]: I0317 16:08:48.266774 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tw9qt"] Mar 17 16:08:48 crc kubenswrapper[4767]: E0317 16:08:48.361615 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda87e499b_99c4_440b_8f1f_2d9e6ef1cd52.slice/crio-72b84b762715964cba4e0b080485b73617ff7e0a0010833944f58f527307b8f1\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda87e499b_99c4_440b_8f1f_2d9e6ef1cd52.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:08:48 crc kubenswrapper[4767]: E0317 16:08:48.363645 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda87e499b_99c4_440b_8f1f_2d9e6ef1cd52.slice/crio-72b84b762715964cba4e0b080485b73617ff7e0a0010833944f58f527307b8f1\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda87e499b_99c4_440b_8f1f_2d9e6ef1cd52.slice\": RecentStats: unable to find data in memory cache]" Mar 17 16:08:48 crc kubenswrapper[4767]: I0317 16:08:48.965768 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d24c721d-4968-477f-ba8e-23f6d31629d2","Type":"ContainerStarted","Data":"c34c899dd966bf1c15b5c7fe589d5c98cf114bdf10d678f6470a8de42dea4ea6"} Mar 17 16:08:48 crc kubenswrapper[4767]: I0317 16:08:48.967772 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 16:08:48 crc kubenswrapper[4767]: I0317 16:08:48.979012 4767 generic.go:334] "Generic (PLEG): container finished" podID="04dde071-2bb2-4e33-8253-d80c863b55e9" containerID="e7db400a2cbbf1b9f8c79681b1cc1aa664ab465a73244e2d7d5fef8109eb9c52" exitCode=0 Mar 17 16:08:48 crc kubenswrapper[4767]: I0317 16:08:48.979108 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" event={"ID":"04dde071-2bb2-4e33-8253-d80c863b55e9","Type":"ContainerDied","Data":"e7db400a2cbbf1b9f8c79681b1cc1aa664ab465a73244e2d7d5fef8109eb9c52"} Mar 17 16:08:48 crc kubenswrapper[4767]: I0317 16:08:48.980452 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" event={"ID":"04dde071-2bb2-4e33-8253-d80c863b55e9","Type":"ContainerStarted","Data":"433a4e8d3aec0543bf9ce3f1b20bed5723a1a4ccc14e08779e02570f225e5874"} Mar 17 16:08:49 crc kubenswrapper[4767]: I0317 16:08:49.025618 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.278901524 podStartE2EDuration="11.025582243s" podCreationTimestamp="2026-03-17 16:08:38 +0000 UTC" firstStartedPulling="2026-03-17 16:08:39.29840936 +0000 UTC m=+1910.711725407" lastFinishedPulling="2026-03-17 16:08:48.045090079 +0000 UTC m=+1919.458406126" observedRunningTime="2026-03-17 16:08:48.999087755 +0000 UTC m=+1920.412403802" watchObservedRunningTime="2026-03-17 16:08:49.025582243 +0000 UTC m=+1920.438898290" Mar 17 16:08:49 crc kubenswrapper[4767]: I0317 16:08:49.056226 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:08:49 crc kubenswrapper[4767]: I0317 16:08:49.126590 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:08:49 crc kubenswrapper[4767]: I0317 16:08:49.378019 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" path="/var/lib/kubelet/pods/a87e499b-99c4-440b-8f1f-2d9e6ef1cd52/volumes" Mar 17 16:08:49 crc kubenswrapper[4767]: I0317 16:08:49.521107 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lc99p"] Mar 17 16:08:49 crc kubenswrapper[4767]: I0317 16:08:49.997380 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" event={"ID":"04dde071-2bb2-4e33-8253-d80c863b55e9","Type":"ContainerStarted","Data":"6d36b601aee2fce655a5a3fc6b5abbbd960032a5b5c1b3867f977d64afa9d367"} Mar 17 16:08:50 crc kubenswrapper[4767]: I0317 16:08:50.037874 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" podStartSLOduration=4.037847956 podStartE2EDuration="4.037847956s" podCreationTimestamp="2026-03-17 16:08:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:08:50.025303697 +0000 UTC m=+1921.438619794" watchObservedRunningTime="2026-03-17 16:08:50.037847956 +0000 UTC m=+1921.451164003" Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.009994 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lc99p" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="registry-server" containerID="cri-o://0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5" gracePeriod=2 Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.013891 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.655577 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.710592 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk9t4\" (UniqueName: \"kubernetes.io/projected/1dea3dfa-5821-4999-99f8-7c17d9ec7515-kube-api-access-pk9t4\") pod \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.710770 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-catalog-content\") pod \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.710937 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-utilities\") pod \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\" (UID: \"1dea3dfa-5821-4999-99f8-7c17d9ec7515\") " Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.712204 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-utilities" (OuterVolumeSpecName: "utilities") pod "1dea3dfa-5821-4999-99f8-7c17d9ec7515" (UID: "1dea3dfa-5821-4999-99f8-7c17d9ec7515"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.716670 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dea3dfa-5821-4999-99f8-7c17d9ec7515-kube-api-access-pk9t4" (OuterVolumeSpecName: "kube-api-access-pk9t4") pod "1dea3dfa-5821-4999-99f8-7c17d9ec7515" (UID: "1dea3dfa-5821-4999-99f8-7c17d9ec7515"). InnerVolumeSpecName "kube-api-access-pk9t4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.814065 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk9t4\" (UniqueName: \"kubernetes.io/projected/1dea3dfa-5821-4999-99f8-7c17d9ec7515-kube-api-access-pk9t4\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.814115 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.847300 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1dea3dfa-5821-4999-99f8-7c17d9ec7515" (UID: "1dea3dfa-5821-4999-99f8-7c17d9ec7515"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:08:51 crc kubenswrapper[4767]: I0317 16:08:51.917745 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dea3dfa-5821-4999-99f8-7c17d9ec7515-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.032110 4767 generic.go:334] "Generic (PLEG): container finished" podID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerID="0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5" exitCode=0 Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.032252 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc99p" Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.032255 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc99p" event={"ID":"1dea3dfa-5821-4999-99f8-7c17d9ec7515","Type":"ContainerDied","Data":"0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5"} Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.032408 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc99p" event={"ID":"1dea3dfa-5821-4999-99f8-7c17d9ec7515","Type":"ContainerDied","Data":"84bbb39b8f9c69ebbb6ee248a2254287c39668677cd66596baf7a934229d8c89"} Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.032444 4767 scope.go:117] "RemoveContainer" containerID="0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5" Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.073504 4767 scope.go:117] "RemoveContainer" containerID="6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a" Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.085805 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lc99p"] Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.100027 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lc99p"] Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.116696 4767 scope.go:117] "RemoveContainer" containerID="adf8ca6066c5497a961566785101b0324f1fe442e2ebcc3db1c6e8d5b3b13044" Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.265464 4767 scope.go:117] "RemoveContainer" containerID="0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5" Mar 17 16:08:52 crc kubenswrapper[4767]: E0317 16:08:52.266095 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5\": container with ID starting with 0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5 not found: ID does not exist" containerID="0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5" Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.266157 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5"} err="failed to get container status \"0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5\": rpc error: code = NotFound desc = could not find container \"0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5\": container with ID starting with 0e79102e203ad14f0b496686295d1be1dd0dcf0c37b72f447d58b736c2bce2d5 not found: ID does not exist" Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.266210 4767 scope.go:117] "RemoveContainer" containerID="6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a" Mar 17 16:08:52 crc kubenswrapper[4767]: E0317 16:08:52.267021 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a\": container with ID starting with 6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a not found: ID does not exist" containerID="6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a" Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.267399 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a"} err="failed to get container status \"6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a\": rpc error: code = NotFound desc = could not find container \"6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a\": container with ID starting with 6fbcb3bc70b0a19f8888ef079ed44abd19a80393e59a89aff1a74f927552e90a not found: ID does not exist" Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.267546 4767 scope.go:117] "RemoveContainer" containerID="adf8ca6066c5497a961566785101b0324f1fe442e2ebcc3db1c6e8d5b3b13044" Mar 17 16:08:52 crc kubenswrapper[4767]: E0317 16:08:52.268150 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adf8ca6066c5497a961566785101b0324f1fe442e2ebcc3db1c6e8d5b3b13044\": container with ID starting with adf8ca6066c5497a961566785101b0324f1fe442e2ebcc3db1c6e8d5b3b13044 not found: ID does not exist" containerID="adf8ca6066c5497a961566785101b0324f1fe442e2ebcc3db1c6e8d5b3b13044" Mar 17 16:08:52 crc kubenswrapper[4767]: I0317 16:08:52.268262 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adf8ca6066c5497a961566785101b0324f1fe442e2ebcc3db1c6e8d5b3b13044"} err="failed to get container status \"adf8ca6066c5497a961566785101b0324f1fe442e2ebcc3db1c6e8d5b3b13044\": rpc error: code = NotFound desc = could not find container \"adf8ca6066c5497a961566785101b0324f1fe442e2ebcc3db1c6e8d5b3b13044\": container with ID starting with adf8ca6066c5497a961566785101b0324f1fe442e2ebcc3db1c6e8d5b3b13044 not found: ID does not exist" Mar 17 16:08:53 crc kubenswrapper[4767]: I0317 16:08:53.049089 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-4zghh" event={"ID":"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3","Type":"ContainerStarted","Data":"da6c402c498e398f77675e3566c0f0e0d201675f910e2c5130652af46c55ad1f"} Mar 17 16:08:53 crc kubenswrapper[4767]: I0317 16:08:53.084086 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-4zghh" podStartSLOduration=1.658197623 podStartE2EDuration="46.084057252s" podCreationTimestamp="2026-03-17 16:08:07 +0000 UTC" firstStartedPulling="2026-03-17 16:08:08.134076962 +0000 UTC m=+1879.547393009" lastFinishedPulling="2026-03-17 16:08:52.559936591 +0000 UTC m=+1923.973252638" observedRunningTime="2026-03-17 16:08:53.073803519 +0000 UTC m=+1924.487119576" watchObservedRunningTime="2026-03-17 16:08:53.084057252 +0000 UTC m=+1924.497373299" Mar 17 16:08:53 crc kubenswrapper[4767]: I0317 16:08:53.369116 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" path="/var/lib/kubelet/pods/1dea3dfa-5821-4999-99f8-7c17d9ec7515/volumes" Mar 17 16:08:55 crc kubenswrapper[4767]: I0317 16:08:55.075145 4767 generic.go:334] "Generic (PLEG): container finished" podID="d28d7ecb-5431-4abd-beb2-f4a4c22df3f3" containerID="da6c402c498e398f77675e3566c0f0e0d201675f910e2c5130652af46c55ad1f" exitCode=0 Mar 17 16:08:55 crc kubenswrapper[4767]: I0317 16:08:55.075223 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-4zghh" event={"ID":"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3","Type":"ContainerDied","Data":"da6c402c498e398f77675e3566c0f0e0d201675f910e2c5130652af46c55ad1f"} Mar 17 16:08:55 crc kubenswrapper[4767]: I0317 16:08:55.356026 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:08:55 crc kubenswrapper[4767]: E0317 16:08:55.356826 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:08:56 crc kubenswrapper[4767]: I0317 16:08:56.640359 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:56 crc kubenswrapper[4767]: I0317 16:08:56.705063 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbxbp\" (UniqueName: \"kubernetes.io/projected/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-kube-api-access-kbxbp\") pod \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " Mar 17 16:08:56 crc kubenswrapper[4767]: I0317 16:08:56.705670 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-combined-ca-bundle\") pod \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " Mar 17 16:08:56 crc kubenswrapper[4767]: I0317 16:08:56.705817 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-config-data\") pod \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\" (UID: \"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3\") " Mar 17 16:08:56 crc kubenswrapper[4767]: I0317 16:08:56.711633 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-kube-api-access-kbxbp" (OuterVolumeSpecName: "kube-api-access-kbxbp") pod "d28d7ecb-5431-4abd-beb2-f4a4c22df3f3" (UID: "d28d7ecb-5431-4abd-beb2-f4a4c22df3f3"). InnerVolumeSpecName "kube-api-access-kbxbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:56 crc kubenswrapper[4767]: I0317 16:08:56.745576 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d28d7ecb-5431-4abd-beb2-f4a4c22df3f3" (UID: "d28d7ecb-5431-4abd-beb2-f4a4c22df3f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:56 crc kubenswrapper[4767]: I0317 16:08:56.809722 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:56 crc kubenswrapper[4767]: I0317 16:08:56.809784 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbxbp\" (UniqueName: \"kubernetes.io/projected/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-kube-api-access-kbxbp\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:56 crc kubenswrapper[4767]: I0317 16:08:56.825725 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-config-data" (OuterVolumeSpecName: "config-data") pod "d28d7ecb-5431-4abd-beb2-f4a4c22df3f3" (UID: "d28d7ecb-5431-4abd-beb2-f4a4c22df3f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:08:56 crc kubenswrapper[4767]: I0317 16:08:56.915887 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:57 crc kubenswrapper[4767]: I0317 16:08:57.100507 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-4zghh" event={"ID":"d28d7ecb-5431-4abd-beb2-f4a4c22df3f3","Type":"ContainerDied","Data":"fc90611c267553f377500baec18af3a5f7e6c01a09aed5bd5c8e43fb5ffc2ce9"} Mar 17 16:08:57 crc kubenswrapper[4767]: I0317 16:08:57.100562 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc90611c267553f377500baec18af3a5f7e6c01a09aed5bd5c8e43fb5ffc2ce9" Mar 17 16:08:57 crc kubenswrapper[4767]: I0317 16:08:57.100627 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-4zghh" Mar 17 16:08:57 crc kubenswrapper[4767]: I0317 16:08:57.285373 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f6df4f56c-jxg2g" Mar 17 16:08:57 crc kubenswrapper[4767]: I0317 16:08:57.380663 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-hm5mf"] Mar 17 16:08:57 crc kubenswrapper[4767]: I0317 16:08:57.380993 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" podUID="a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" containerName="dnsmasq-dns" containerID="cri-o://95da0a180a4468719b26f8b482508ee25c55c837077e708b354b1bcc56c9edb7" gracePeriod=10 Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.145997 4767 generic.go:334] "Generic (PLEG): container finished" podID="a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" containerID="95da0a180a4468719b26f8b482508ee25c55c837077e708b354b1bcc56c9edb7" exitCode=0 Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.146358 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" event={"ID":"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba","Type":"ContainerDied","Data":"95da0a180a4468719b26f8b482508ee25c55c837077e708b354b1bcc56c9edb7"} Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.146399 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" event={"ID":"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba","Type":"ContainerDied","Data":"72e6558b1fd18e7c0bc1b407e59139269635f6f0ce483cd8296093c17fc7ca66"} Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.146416 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72e6558b1fd18e7c0bc1b407e59139269635f6f0ce483cd8296093c17fc7ca66" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.151705 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.268704 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-config\") pod \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.268962 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-openstack-edpm-ipam\") pod \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.269005 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pztkq\" (UniqueName: \"kubernetes.io/projected/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-kube-api-access-pztkq\") pod \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.269036 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-sb\") pod \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.269115 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-swift-storage-0\") pod \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.269201 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-nb\") pod \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.269265 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-svc\") pod \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\" (UID: \"a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba\") " Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.286510 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-kube-api-access-pztkq" (OuterVolumeSpecName: "kube-api-access-pztkq") pod "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" (UID: "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba"). InnerVolumeSpecName "kube-api-access-pztkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.353277 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-config" (OuterVolumeSpecName: "config") pod "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" (UID: "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.357769 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" (UID: "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.367237 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" (UID: "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.372358 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.372423 4767 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.372435 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-config\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.372451 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pztkq\" (UniqueName: \"kubernetes.io/projected/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-kube-api-access-pztkq\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.390649 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" (UID: "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.397785 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" (UID: "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.402255 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" (UID: "a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.479714 4767 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.479754 4767 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.479765 4767 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.960387 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7fb4b69657-nxpx7"] Mar 17 16:08:58 crc kubenswrapper[4767]: E0317 16:08:58.961339 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28d7ecb-5431-4abd-beb2-f4a4c22df3f3" containerName="heat-db-sync" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.961369 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28d7ecb-5431-4abd-beb2-f4a4c22df3f3" containerName="heat-db-sync" Mar 17 16:08:58 crc kubenswrapper[4767]: E0317 16:08:58.961400 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" containerName="init" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.961410 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" containerName="init" Mar 17 16:08:58 crc kubenswrapper[4767]: E0317 16:08:58.961422 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="extract-content" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.961432 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="extract-content" Mar 17 16:08:58 crc kubenswrapper[4767]: E0317 16:08:58.961465 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="extract-utilities" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.961475 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="extract-utilities" Mar 17 16:08:58 crc kubenswrapper[4767]: E0317 16:08:58.961493 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="registry-server" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.961501 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="registry-server" Mar 17 16:08:58 crc kubenswrapper[4767]: E0317 16:08:58.961542 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" containerName="dnsmasq-dns" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.961551 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" containerName="dnsmasq-dns" Mar 17 16:08:58 crc kubenswrapper[4767]: E0317 16:08:58.961565 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" containerName="dnsmasq-dns" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.961573 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" containerName="dnsmasq-dns" Mar 17 16:08:58 crc kubenswrapper[4767]: E0317 16:08:58.961607 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" containerName="init" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.961619 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" containerName="init" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.961952 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dea3dfa-5821-4999-99f8-7c17d9ec7515" containerName="registry-server" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.961992 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" containerName="dnsmasq-dns" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.962012 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a87e499b-99c4-440b-8f1f-2d9e6ef1cd52" containerName="dnsmasq-dns" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.962040 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d28d7ecb-5431-4abd-beb2-f4a4c22df3f3" containerName="heat-db-sync" Mar 17 16:08:58 crc kubenswrapper[4767]: I0317 16:08:58.963471 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.007132 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7fb4b69657-nxpx7"] Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.077284 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-69598cd69b-zjhkv"] Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.079397 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.101821 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/35c67f99-377a-4787-bf78-fe50d478eadd-config-data-custom\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.101942 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c67f99-377a-4787-bf78-fe50d478eadd-config-data\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.101991 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzhps\" (UniqueName: \"kubernetes.io/projected/35c67f99-377a-4787-bf78-fe50d478eadd-kube-api-access-nzhps\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.102029 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c67f99-377a-4787-bf78-fe50d478eadd-combined-ca-bundle\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.111918 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69598cd69b-zjhkv"] Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.137835 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5b66464f74-f6tl4"] Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.139932 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.153161 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5b66464f74-f6tl4"] Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.163745 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-hm5mf" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.205121 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c67f99-377a-4787-bf78-fe50d478eadd-combined-ca-bundle\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.205203 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-combined-ca-bundle\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.205310 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5m74\" (UniqueName: \"kubernetes.io/projected/a8879e45-8cce-4768-a60d-e873a344d76b-kube-api-access-z5m74\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.205380 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-internal-tls-certs\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.205428 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/35c67f99-377a-4787-bf78-fe50d478eadd-config-data-custom\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.205458 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-config-data-custom\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.205537 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c67f99-377a-4787-bf78-fe50d478eadd-config-data\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.205576 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-config-data\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.205597 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-public-tls-certs\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.205634 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzhps\" (UniqueName: \"kubernetes.io/projected/35c67f99-377a-4787-bf78-fe50d478eadd-kube-api-access-nzhps\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.210319 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-hm5mf"] Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.211776 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c67f99-377a-4787-bf78-fe50d478eadd-combined-ca-bundle\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.211985 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/35c67f99-377a-4787-bf78-fe50d478eadd-config-data-custom\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.212084 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c67f99-377a-4787-bf78-fe50d478eadd-config-data\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.224802 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-hm5mf"] Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.232670 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzhps\" (UniqueName: \"kubernetes.io/projected/35c67f99-377a-4787-bf78-fe50d478eadd-kube-api-access-nzhps\") pod \"heat-engine-7fb4b69657-nxpx7\" (UID: \"35c67f99-377a-4787-bf78-fe50d478eadd\") " pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.304992 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.308058 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn479\" (UniqueName: \"kubernetes.io/projected/38a99155-2272-4a36-ba8d-a04be51218af-kube-api-access-zn479\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.308196 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-public-tls-certs\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.308266 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5m74\" (UniqueName: \"kubernetes.io/projected/a8879e45-8cce-4768-a60d-e873a344d76b-kube-api-access-z5m74\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.308326 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-config-data\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.308363 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-internal-tls-certs\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.308422 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-internal-tls-certs\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.308499 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-config-data-custom\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.308574 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-combined-ca-bundle\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.311735 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-config-data-custom\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.311794 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-config-data\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.311827 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-public-tls-certs\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.312007 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-combined-ca-bundle\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.317569 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-internal-tls-certs\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.321141 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-config-data-custom\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.324095 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-combined-ca-bundle\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.327279 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-public-tls-certs\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.330321 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8879e45-8cce-4768-a60d-e873a344d76b-config-data\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.343924 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5m74\" (UniqueName: \"kubernetes.io/projected/a8879e45-8cce-4768-a60d-e873a344d76b-kube-api-access-z5m74\") pod \"heat-api-69598cd69b-zjhkv\" (UID: \"a8879e45-8cce-4768-a60d-e873a344d76b\") " pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.396865 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba" path="/var/lib/kubelet/pods/a40b3eb6-24e8-4aa2-bbbb-ae1233efcbba/volumes" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.414910 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-internal-tls-certs\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.415029 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-combined-ca-bundle\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.415102 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-config-data-custom\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.415213 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn479\" (UniqueName: \"kubernetes.io/projected/38a99155-2272-4a36-ba8d-a04be51218af-kube-api-access-zn479\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.415258 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-public-tls-certs\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.415302 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-config-data\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.419296 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-config-data-custom\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.419680 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-internal-tls-certs\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.420152 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.420622 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-config-data\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.424931 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-public-tls-certs\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.427684 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38a99155-2272-4a36-ba8d-a04be51218af-combined-ca-bundle\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.446382 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn479\" (UniqueName: \"kubernetes.io/projected/38a99155-2272-4a36-ba8d-a04be51218af-kube-api-access-zn479\") pod \"heat-cfnapi-5b66464f74-f6tl4\" (UID: \"38a99155-2272-4a36-ba8d-a04be51218af\") " pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.462554 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:08:59 crc kubenswrapper[4767]: I0317 16:08:59.910344 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7fb4b69657-nxpx7"] Mar 17 16:09:00 crc kubenswrapper[4767]: I0317 16:09:00.081867 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69598cd69b-zjhkv"] Mar 17 16:09:00 crc kubenswrapper[4767]: W0317 16:09:00.098067 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8879e45_8cce_4768_a60d_e873a344d76b.slice/crio-f4812c2fa0e2182c461a394034745fe7bfdff456e2a30f901854b9ad36fad255 WatchSource:0}: Error finding container f4812c2fa0e2182c461a394034745fe7bfdff456e2a30f901854b9ad36fad255: Status 404 returned error can't find the container with id f4812c2fa0e2182c461a394034745fe7bfdff456e2a30f901854b9ad36fad255 Mar 17 16:09:00 crc kubenswrapper[4767]: I0317 16:09:00.183751 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7fb4b69657-nxpx7" event={"ID":"35c67f99-377a-4787-bf78-fe50d478eadd","Type":"ContainerStarted","Data":"a7e36df16ce31e9b70ff274dbde10f0b70b71c091aaa72ee6d93586dd099f5ee"} Mar 17 16:09:00 crc kubenswrapper[4767]: I0317 16:09:00.185928 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69598cd69b-zjhkv" event={"ID":"a8879e45-8cce-4768-a60d-e873a344d76b","Type":"ContainerStarted","Data":"f4812c2fa0e2182c461a394034745fe7bfdff456e2a30f901854b9ad36fad255"} Mar 17 16:09:00 crc kubenswrapper[4767]: I0317 16:09:00.258886 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5b66464f74-f6tl4"] Mar 17 16:09:01 crc kubenswrapper[4767]: I0317 16:09:01.207218 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5b66464f74-f6tl4" event={"ID":"38a99155-2272-4a36-ba8d-a04be51218af","Type":"ContainerStarted","Data":"73567b7de2984ef344ac9a6c1ba7022b2b5c294500483b81b5e22b73e12c0cd8"} Mar 17 16:09:01 crc kubenswrapper[4767]: I0317 16:09:01.215425 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7fb4b69657-nxpx7" event={"ID":"35c67f99-377a-4787-bf78-fe50d478eadd","Type":"ContainerStarted","Data":"0c05f2e3fde56880224dd9fb8647478836d3221316c9cfd14330e26c03fb4dd6"} Mar 17 16:09:01 crc kubenswrapper[4767]: I0317 16:09:01.217810 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:09:01 crc kubenswrapper[4767]: I0317 16:09:01.253410 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7fb4b69657-nxpx7" podStartSLOduration=3.253381267 podStartE2EDuration="3.253381267s" podCreationTimestamp="2026-03-17 16:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:09:01.23810181 +0000 UTC m=+1932.651417867" watchObservedRunningTime="2026-03-17 16:09:01.253381267 +0000 UTC m=+1932.666697314" Mar 17 16:09:03 crc kubenswrapper[4767]: I0317 16:09:03.273120 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5b66464f74-f6tl4" event={"ID":"38a99155-2272-4a36-ba8d-a04be51218af","Type":"ContainerStarted","Data":"ec2d321c9941d2a02c88c39b54761f422ac324b756e152d4a3b35633b3386d36"} Mar 17 16:09:03 crc kubenswrapper[4767]: I0317 16:09:03.274407 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:09:03 crc kubenswrapper[4767]: I0317 16:09:03.284309 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69598cd69b-zjhkv" event={"ID":"a8879e45-8cce-4768-a60d-e873a344d76b","Type":"ContainerStarted","Data":"704e612e44ce31ae742dd46abb56bc4c08ff91155fd0fb438a7c1bfb3e668a66"} Mar 17 16:09:03 crc kubenswrapper[4767]: I0317 16:09:03.320112 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5b66464f74-f6tl4" podStartSLOduration=2.682354806 podStartE2EDuration="4.320085958s" podCreationTimestamp="2026-03-17 16:08:59 +0000 UTC" firstStartedPulling="2026-03-17 16:09:00.257892945 +0000 UTC m=+1931.671208992" lastFinishedPulling="2026-03-17 16:09:01.895624097 +0000 UTC m=+1933.308940144" observedRunningTime="2026-03-17 16:09:03.298905722 +0000 UTC m=+1934.712221809" watchObservedRunningTime="2026-03-17 16:09:03.320085958 +0000 UTC m=+1934.733402005" Mar 17 16:09:03 crc kubenswrapper[4767]: I0317 16:09:03.341162 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-69598cd69b-zjhkv" podStartSLOduration=2.548681743 podStartE2EDuration="4.34113559s" podCreationTimestamp="2026-03-17 16:08:59 +0000 UTC" firstStartedPulling="2026-03-17 16:09:00.101471901 +0000 UTC m=+1931.514787948" lastFinishedPulling="2026-03-17 16:09:01.893925748 +0000 UTC m=+1933.307241795" observedRunningTime="2026-03-17 16:09:03.333648236 +0000 UTC m=+1934.746964293" watchObservedRunningTime="2026-03-17 16:09:03.34113559 +0000 UTC m=+1934.754451637" Mar 17 16:09:04 crc kubenswrapper[4767]: I0317 16:09:04.295543 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:09:06 crc kubenswrapper[4767]: I0317 16:09:06.695495 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bfd60c23-044e-4fae-8289-c460e98160fb" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.1.22:3000/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 16:09:08 crc kubenswrapper[4767]: I0317 16:09:08.354959 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:09:08 crc kubenswrapper[4767]: I0317 16:09:08.509401 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 17 16:09:09 crc kubenswrapper[4767]: I0317 16:09:09.374508 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"66ab96d70d85b8ad4af97b2cc9506406fc51b53c84a92e8d73017d6fc8cc0c7f"} Mar 17 16:09:11 crc kubenswrapper[4767]: I0317 16:09:11.141302 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-69598cd69b-zjhkv" Mar 17 16:09:11 crc kubenswrapper[4767]: I0317 16:09:11.238467 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6449cd4787-bslls"] Mar 17 16:09:11 crc kubenswrapper[4767]: I0317 16:09:11.238736 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-6449cd4787-bslls" podUID="6fdb5a6c-03a0-4184-8f33-3b7244792cc5" containerName="heat-api" containerID="cri-o://e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670" gracePeriod=60 Mar 17 16:09:11 crc kubenswrapper[4767]: I0317 16:09:11.741038 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5b66464f74-f6tl4" Mar 17 16:09:11 crc kubenswrapper[4767]: I0317 16:09:11.863245 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-56f748dcbf-9x7mb"] Mar 17 16:09:11 crc kubenswrapper[4767]: I0317 16:09:11.864285 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" podUID="43b3b246-f5d3-4311-ab67-d30e53ac7082" containerName="heat-cfnapi" containerID="cri-o://47cdf9affcddbe0aa7ad601baeab2ee25ee63588a6b2bca9a3c0ee67e20f0c0d" gracePeriod=60 Mar 17 16:09:14 crc kubenswrapper[4767]: I0317 16:09:14.468842 4767 generic.go:334] "Generic (PLEG): container finished" podID="a298936e-5022-44de-9d7a-65f67ec76087" containerID="4338209039d729077a65c9693399c1133c96fd0b7fa40d3a483955d6e2e4f031" exitCode=0 Mar 17 16:09:14 crc kubenswrapper[4767]: I0317 16:09:14.469524 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a298936e-5022-44de-9d7a-65f67ec76087","Type":"ContainerDied","Data":"4338209039d729077a65c9693399c1133c96fd0b7fa40d3a483955d6e2e4f031"} Mar 17 16:09:14 crc kubenswrapper[4767]: I0317 16:09:14.479104 4767 generic.go:334] "Generic (PLEG): container finished" podID="cce1c59c-7c85-4922-bbbc-462261019bc9" containerID="7d505a0d7bbd0bd3e6942d51090cd0e1bfebadb499cf0b186996aa2a0c474706" exitCode=0 Mar 17 16:09:14 crc kubenswrapper[4767]: I0317 16:09:14.479165 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cce1c59c-7c85-4922-bbbc-462261019bc9","Type":"ContainerDied","Data":"7d505a0d7bbd0bd3e6942d51090cd0e1bfebadb499cf0b186996aa2a0c474706"} Mar 17 16:09:14 crc kubenswrapper[4767]: I0317 16:09:14.669166 4767 scope.go:117] "RemoveContainer" containerID="0dab081d8810ac517e5792d0ceeee372d5ab8873ed4c7434b2d8caee8d2deda1" Mar 17 16:09:14 crc kubenswrapper[4767]: I0317 16:09:14.718001 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-6449cd4787-bslls" podUID="6fdb5a6c-03a0-4184-8f33-3b7244792cc5" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.230:8004/healthcheck\": read tcp 10.217.0.2:42972->10.217.0.230:8004: read: connection reset by peer" Mar 17 16:09:14 crc kubenswrapper[4767]: I0317 16:09:14.894743 4767 scope.go:117] "RemoveContainer" containerID="218dbac6c2516ec6291a0108ea2ab138bafea03e1445ea1e37fdbff853f3e8ac" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.050761 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" podUID="43b3b246-f5d3-4311-ab67-d30e53ac7082" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.231:8000/healthcheck\": read tcp 10.217.0.2:38736->10.217.0.231:8000: read: connection reset by peer" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.330755 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.370814 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data-custom\") pod \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.370957 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-public-tls-certs\") pod \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.370995 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjqt8\" (UniqueName: \"kubernetes.io/projected/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-kube-api-access-gjqt8\") pod \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.373388 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-combined-ca-bundle\") pod \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.373538 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-internal-tls-certs\") pod \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.373605 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data\") pod \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\" (UID: \"6fdb5a6c-03a0-4184-8f33-3b7244792cc5\") " Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.582810 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6fdb5a6c-03a0-4184-8f33-3b7244792cc5" (UID: "6fdb5a6c-03a0-4184-8f33-3b7244792cc5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.607503 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-kube-api-access-gjqt8" (OuterVolumeSpecName: "kube-api-access-gjqt8") pod "6fdb5a6c-03a0-4184-8f33-3b7244792cc5" (UID: "6fdb5a6c-03a0-4184-8f33-3b7244792cc5"). InnerVolumeSpecName "kube-api-access-gjqt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.675323 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6fdb5a6c-03a0-4184-8f33-3b7244792cc5" (UID: "6fdb5a6c-03a0-4184-8f33-3b7244792cc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.720595 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6fdb5a6c-03a0-4184-8f33-3b7244792cc5" (UID: "6fdb5a6c-03a0-4184-8f33-3b7244792cc5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.817069 4767 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.817116 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.817135 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjqt8\" (UniqueName: \"kubernetes.io/projected/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-kube-api-access-gjqt8\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.817151 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.856362 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data" (OuterVolumeSpecName: "config-data") pod "6fdb5a6c-03a0-4184-8f33-3b7244792cc5" (UID: "6fdb5a6c-03a0-4184-8f33-3b7244792cc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.920149 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.923385 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6fdb5a6c-03a0-4184-8f33-3b7244792cc5" (UID: "6fdb5a6c-03a0-4184-8f33-3b7244792cc5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.930121 4767 generic.go:334] "Generic (PLEG): container finished" podID="6fdb5a6c-03a0-4184-8f33-3b7244792cc5" containerID="e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670" exitCode=0 Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.930154 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6449cd4787-bslls" event={"ID":"6fdb5a6c-03a0-4184-8f33-3b7244792cc5","Type":"ContainerDied","Data":"e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670"} Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.964199 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6449cd4787-bslls" event={"ID":"6fdb5a6c-03a0-4184-8f33-3b7244792cc5","Type":"ContainerDied","Data":"14fd68462401f4b9e9a32f3322499914628008b9f9298e9a1b777c9d70ad6e84"} Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.964269 4767 scope.go:117] "RemoveContainer" containerID="e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.930242 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6449cd4787-bslls" Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.976411 4767 generic.go:334] "Generic (PLEG): container finished" podID="43b3b246-f5d3-4311-ab67-d30e53ac7082" containerID="47cdf9affcddbe0aa7ad601baeab2ee25ee63588a6b2bca9a3c0ee67e20f0c0d" exitCode=0 Mar 17 16:09:15 crc kubenswrapper[4767]: I0317 16:09:15.976503 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" event={"ID":"43b3b246-f5d3-4311-ab67-d30e53ac7082","Type":"ContainerDied","Data":"47cdf9affcddbe0aa7ad601baeab2ee25ee63588a6b2bca9a3c0ee67e20f0c0d"} Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.003696 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.006691 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a298936e-5022-44de-9d7a-65f67ec76087","Type":"ContainerStarted","Data":"7211c276260f751018bbeb4b2628e1606f1caffa2471edced15866ffe13cb9bc"} Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.009085 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-2" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.021431 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cce1c59c-7c85-4922-bbbc-462261019bc9","Type":"ContainerStarted","Data":"a4bdb16e762a45589cd62d18a9608eb5146e0aa739829970dc0a598a3960e4c8"} Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.024436 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.038691 4767 scope.go:117] "RemoveContainer" containerID="e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670" Mar 17 16:09:16 crc kubenswrapper[4767]: E0317 16:09:16.039776 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670\": container with ID starting with e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670 not found: ID does not exist" containerID="e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.039827 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670"} err="failed to get container status \"e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670\": rpc error: code = NotFound desc = could not find container \"e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670\": container with ID starting with e654adb59d3cd1a2add2875b15735eb83c32314ff9a68ff82268394bb6dc6670 not found: ID does not exist" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.045394 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6449cd4787-bslls"] Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.064328 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6449cd4787-bslls"] Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.066718 4767 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdb5a6c-03a0-4184-8f33-3b7244792cc5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.069799 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-2" podStartSLOduration=45.069764468 podStartE2EDuration="45.069764468s" podCreationTimestamp="2026-03-17 16:08:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:09:16.060883025 +0000 UTC m=+1947.474199072" watchObservedRunningTime="2026-03-17 16:09:16.069764468 +0000 UTC m=+1947.483080515" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.155049 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.155024503 podStartE2EDuration="38.155024503s" podCreationTimestamp="2026-03-17 16:08:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:09:16.120747698 +0000 UTC m=+1947.534063755" watchObservedRunningTime="2026-03-17 16:09:16.155024503 +0000 UTC m=+1947.568340550" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.170539 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-internal-tls-certs\") pod \"43b3b246-f5d3-4311-ab67-d30e53ac7082\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.170615 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-combined-ca-bundle\") pod \"43b3b246-f5d3-4311-ab67-d30e53ac7082\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.170698 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data\") pod \"43b3b246-f5d3-4311-ab67-d30e53ac7082\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.170785 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-public-tls-certs\") pod \"43b3b246-f5d3-4311-ab67-d30e53ac7082\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.170836 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data-custom\") pod \"43b3b246-f5d3-4311-ab67-d30e53ac7082\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.171063 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gt24\" (UniqueName: \"kubernetes.io/projected/43b3b246-f5d3-4311-ab67-d30e53ac7082-kube-api-access-2gt24\") pod \"43b3b246-f5d3-4311-ab67-d30e53ac7082\" (UID: \"43b3b246-f5d3-4311-ab67-d30e53ac7082\") " Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.180785 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43b3b246-f5d3-4311-ab67-d30e53ac7082-kube-api-access-2gt24" (OuterVolumeSpecName: "kube-api-access-2gt24") pod "43b3b246-f5d3-4311-ab67-d30e53ac7082" (UID: "43b3b246-f5d3-4311-ab67-d30e53ac7082"). InnerVolumeSpecName "kube-api-access-2gt24". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.231342 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "43b3b246-f5d3-4311-ab67-d30e53ac7082" (UID: "43b3b246-f5d3-4311-ab67-d30e53ac7082"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.276887 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gt24\" (UniqueName: \"kubernetes.io/projected/43b3b246-f5d3-4311-ab67-d30e53ac7082-kube-api-access-2gt24\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.276927 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.277145 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "43b3b246-f5d3-4311-ab67-d30e53ac7082" (UID: "43b3b246-f5d3-4311-ab67-d30e53ac7082"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.281408 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "43b3b246-f5d3-4311-ab67-d30e53ac7082" (UID: "43b3b246-f5d3-4311-ab67-d30e53ac7082"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.306492 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43b3b246-f5d3-4311-ab67-d30e53ac7082" (UID: "43b3b246-f5d3-4311-ab67-d30e53ac7082"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.310860 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data" (OuterVolumeSpecName: "config-data") pod "43b3b246-f5d3-4311-ab67-d30e53ac7082" (UID: "43b3b246-f5d3-4311-ab67-d30e53ac7082"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.380458 4767 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.380502 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.380514 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.380522 4767 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b3b246-f5d3-4311-ab67-d30e53ac7082-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.563311 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5"] Mar 17 16:09:16 crc kubenswrapper[4767]: E0317 16:09:16.564396 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fdb5a6c-03a0-4184-8f33-3b7244792cc5" containerName="heat-api" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.564415 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fdb5a6c-03a0-4184-8f33-3b7244792cc5" containerName="heat-api" Mar 17 16:09:16 crc kubenswrapper[4767]: E0317 16:09:16.564434 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43b3b246-f5d3-4311-ab67-d30e53ac7082" containerName="heat-cfnapi" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.564441 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="43b3b246-f5d3-4311-ab67-d30e53ac7082" containerName="heat-cfnapi" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.564728 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="43b3b246-f5d3-4311-ab67-d30e53ac7082" containerName="heat-cfnapi" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.564743 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fdb5a6c-03a0-4184-8f33-3b7244792cc5" containerName="heat-api" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.565747 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.570186 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.570791 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.570815 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.571139 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.598902 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5"] Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.699676 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.700180 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjrl7\" (UniqueName: \"kubernetes.io/projected/bec306cb-266c-4161-86cc-da1f1cb66faa-kube-api-access-hjrl7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.700393 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.700469 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.802770 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.803022 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjrl7\" (UniqueName: \"kubernetes.io/projected/bec306cb-266c-4161-86cc-da1f1cb66faa-kube-api-access-hjrl7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.803116 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.803160 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.810308 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.810329 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.816258 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.830696 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjrl7\" (UniqueName: \"kubernetes.io/projected/bec306cb-266c-4161-86cc-da1f1cb66faa-kube-api-access-hjrl7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:16 crc kubenswrapper[4767]: I0317 16:09:16.891098 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:17 crc kubenswrapper[4767]: I0317 16:09:17.053675 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" event={"ID":"43b3b246-f5d3-4311-ab67-d30e53ac7082","Type":"ContainerDied","Data":"08e43ff110b79b14aa80a0d973f9c845c33c453a241ce7b8746fd250b01acbd4"} Mar 17 16:09:17 crc kubenswrapper[4767]: I0317 16:09:17.053748 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56f748dcbf-9x7mb" Mar 17 16:09:17 crc kubenswrapper[4767]: I0317 16:09:17.054030 4767 scope.go:117] "RemoveContainer" containerID="47cdf9affcddbe0aa7ad601baeab2ee25ee63588a6b2bca9a3c0ee67e20f0c0d" Mar 17 16:09:17 crc kubenswrapper[4767]: I0317 16:09:17.137339 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-56f748dcbf-9x7mb"] Mar 17 16:09:17 crc kubenswrapper[4767]: I0317 16:09:17.162905 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-56f748dcbf-9x7mb"] Mar 17 16:09:17 crc kubenswrapper[4767]: I0317 16:09:17.372122 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43b3b246-f5d3-4311-ab67-d30e53ac7082" path="/var/lib/kubelet/pods/43b3b246-f5d3-4311-ab67-d30e53ac7082/volumes" Mar 17 16:09:17 crc kubenswrapper[4767]: I0317 16:09:17.372751 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fdb5a6c-03a0-4184-8f33-3b7244792cc5" path="/var/lib/kubelet/pods/6fdb5a6c-03a0-4184-8f33-3b7244792cc5/volumes" Mar 17 16:09:17 crc kubenswrapper[4767]: I0317 16:09:17.732072 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5"] Mar 17 16:09:18 crc kubenswrapper[4767]: I0317 16:09:18.072819 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" event={"ID":"bec306cb-266c-4161-86cc-da1f1cb66faa","Type":"ContainerStarted","Data":"87ce1dbca28c3a54aad9dcbc761887a5790c2a4510204b2e1c6a1c340acb6b12"} Mar 17 16:09:19 crc kubenswrapper[4767]: I0317 16:09:19.352061 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7fb4b69657-nxpx7" Mar 17 16:09:19 crc kubenswrapper[4767]: I0317 16:09:19.447260 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5bdf466bd4-x9rf5"] Mar 17 16:09:19 crc kubenswrapper[4767]: I0317 16:09:19.447777 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-5bdf466bd4-x9rf5" podUID="16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" containerName="heat-engine" containerID="cri-o://c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8" gracePeriod=60 Mar 17 16:09:20 crc kubenswrapper[4767]: E0317 16:09:20.118843 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 17 16:09:20 crc kubenswrapper[4767]: E0317 16:09:20.127655 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 17 16:09:20 crc kubenswrapper[4767]: E0317 16:09:20.137034 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 17 16:09:20 crc kubenswrapper[4767]: E0317 16:09:20.138829 4767 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-5bdf466bd4-x9rf5" podUID="16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" containerName="heat-engine" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.542429 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-bj7s2"] Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.557100 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-bj7s2"] Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.663131 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-c9h7f"] Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.665164 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.668891 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.699639 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-c9h7f"] Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.758328 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-combined-ca-bundle\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.758386 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fprnd\" (UniqueName: \"kubernetes.io/projected/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-kube-api-access-fprnd\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.758419 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-config-data\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.758560 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-scripts\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.864012 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-combined-ca-bundle\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.864097 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fprnd\" (UniqueName: \"kubernetes.io/projected/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-kube-api-access-fprnd\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.864138 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-config-data\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.864198 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-scripts\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.871748 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-config-data\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.881105 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-scripts\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.881121 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-combined-ca-bundle\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.884341 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fprnd\" (UniqueName: \"kubernetes.io/projected/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-kube-api-access-fprnd\") pod \"aodh-db-sync-c9h7f\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:27 crc kubenswrapper[4767]: I0317 16:09:27.995188 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:28 crc kubenswrapper[4767]: I0317 16:09:28.544548 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 17 16:09:29 crc kubenswrapper[4767]: I0317 16:09:29.368679 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d52c8d8-4951-43d6-8889-60f379933fa7" path="/var/lib/kubelet/pods/9d52c8d8-4951-43d6-8889-60f379933fa7/volumes" Mar 17 16:09:30 crc kubenswrapper[4767]: E0317 16:09:30.116952 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 17 16:09:30 crc kubenswrapper[4767]: E0317 16:09:30.122959 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 17 16:09:30 crc kubenswrapper[4767]: E0317 16:09:30.126603 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 17 16:09:30 crc kubenswrapper[4767]: E0317 16:09:30.126701 4767 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-5bdf466bd4-x9rf5" podUID="16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" containerName="heat-engine" Mar 17 16:09:31 crc kubenswrapper[4767]: I0317 16:09:31.303938 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" event={"ID":"bec306cb-266c-4161-86cc-da1f1cb66faa","Type":"ContainerStarted","Data":"49ac648ab09d6b880b8ffd37021fa56228b4fab88173748f63e9206a7b9bded0"} Mar 17 16:09:31 crc kubenswrapper[4767]: I0317 16:09:31.342881 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-c9h7f"] Mar 17 16:09:31 crc kubenswrapper[4767]: I0317 16:09:31.343958 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" podStartSLOduration=2.284657681 podStartE2EDuration="15.34394114s" podCreationTimestamp="2026-03-17 16:09:16 +0000 UTC" firstStartedPulling="2026-03-17 16:09:17.735094759 +0000 UTC m=+1949.148410806" lastFinishedPulling="2026-03-17 16:09:30.794378218 +0000 UTC m=+1962.207694265" observedRunningTime="2026-03-17 16:09:31.321455061 +0000 UTC m=+1962.734771158" watchObservedRunningTime="2026-03-17 16:09:31.34394114 +0000 UTC m=+1962.757257187" Mar 17 16:09:31 crc kubenswrapper[4767]: W0317 16:09:31.351502 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdaa047c_3098_4ee4_b6bf_ffadc6d6fcb2.slice/crio-f5774996fbd96ea283cb2eb306d53e49a0efe0c4eb15da2da4ec44ace667db5c WatchSource:0}: Error finding container f5774996fbd96ea283cb2eb306d53e49a0efe0c4eb15da2da4ec44ace667db5c: Status 404 returned error can't find the container with id f5774996fbd96ea283cb2eb306d53e49a0efe0c4eb15da2da4ec44ace667db5c Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.107121 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.214624 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86lfx\" (UniqueName: \"kubernetes.io/projected/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-kube-api-access-86lfx\") pod \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.215197 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data\") pod \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.215229 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-combined-ca-bundle\") pod \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.215276 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data-custom\") pod \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\" (UID: \"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6\") " Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.228752 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" (UID: "16c87dcb-97fe-48d4-82b3-fa36fb9c89c6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.229437 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-kube-api-access-86lfx" (OuterVolumeSpecName: "kube-api-access-86lfx") pod "16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" (UID: "16c87dcb-97fe-48d4-82b3-fa36fb9c89c6"). InnerVolumeSpecName "kube-api-access-86lfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.266637 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" (UID: "16c87dcb-97fe-48d4-82b3-fa36fb9c89c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.311844 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data" (OuterVolumeSpecName: "config-data") pod "16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" (UID: "16c87dcb-97fe-48d4-82b3-fa36fb9c89c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.319362 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86lfx\" (UniqueName: \"kubernetes.io/projected/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-kube-api-access-86lfx\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.319395 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.319407 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.319748 4767 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.421376 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-2" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.422069 4767 generic.go:334] "Generic (PLEG): container finished" podID="16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" containerID="c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8" exitCode=0 Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.422183 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5bdf466bd4-x9rf5" event={"ID":"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6","Type":"ContainerDied","Data":"c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8"} Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.422222 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5bdf466bd4-x9rf5" event={"ID":"16c87dcb-97fe-48d4-82b3-fa36fb9c89c6","Type":"ContainerDied","Data":"afba0c2425cec5df41454c84fdd531b5dc53d4cdfd2ba651aa15254f7532244b"} Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.422242 4767 scope.go:117] "RemoveContainer" containerID="c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.422380 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5bdf466bd4-x9rf5" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.434808 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-c9h7f" event={"ID":"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2","Type":"ContainerStarted","Data":"f5774996fbd96ea283cb2eb306d53e49a0efe0c4eb15da2da4ec44ace667db5c"} Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.549450 4767 scope.go:117] "RemoveContainer" containerID="c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8" Mar 17 16:09:32 crc kubenswrapper[4767]: E0317 16:09:32.550124 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8\": container with ID starting with c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8 not found: ID does not exist" containerID="c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.550182 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8"} err="failed to get container status \"c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8\": rpc error: code = NotFound desc = could not find container \"c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8\": container with ID starting with c1715d11eb5b003704a64285df983b5d4b432fcbc8d7f7384c418873b060f4a8 not found: ID does not exist" Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.640003 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5bdf466bd4-x9rf5"] Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.663044 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-5bdf466bd4-x9rf5"] Mar 17 16:09:32 crc kubenswrapper[4767]: I0317 16:09:32.690114 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 17 16:09:33 crc kubenswrapper[4767]: I0317 16:09:33.373476 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" path="/var/lib/kubelet/pods/16c87dcb-97fe-48d4-82b3-fa36fb9c89c6/volumes" Mar 17 16:09:36 crc kubenswrapper[4767]: I0317 16:09:36.498792 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-c9h7f" event={"ID":"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2","Type":"ContainerStarted","Data":"83f91c392163db2a58f38aab200eec75baf8e6ac41502ce4b37edadf01556977"} Mar 17 16:09:36 crc kubenswrapper[4767]: I0317 16:09:36.520866 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-c9h7f" podStartSLOduration=4.799997164 podStartE2EDuration="9.520842661s" podCreationTimestamp="2026-03-17 16:09:27 +0000 UTC" firstStartedPulling="2026-03-17 16:09:31.354524352 +0000 UTC m=+1962.767840399" lastFinishedPulling="2026-03-17 16:09:36.075369849 +0000 UTC m=+1967.488685896" observedRunningTime="2026-03-17 16:09:36.514617354 +0000 UTC m=+1967.927933411" watchObservedRunningTime="2026-03-17 16:09:36.520842661 +0000 UTC m=+1967.934158708" Mar 17 16:09:38 crc kubenswrapper[4767]: I0317 16:09:38.031034 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-1" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" containerName="rabbitmq" containerID="cri-o://2f570a2d5d87e91310496d39d3f5743a020b78ef2da12089bcec9dc84e6c63c1" gracePeriod=604795 Mar 17 16:09:39 crc kubenswrapper[4767]: I0317 16:09:39.349116 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.133:5671: connect: connection refused" Mar 17 16:09:39 crc kubenswrapper[4767]: I0317 16:09:39.538885 4767 generic.go:334] "Generic (PLEG): container finished" podID="fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2" containerID="83f91c392163db2a58f38aab200eec75baf8e6ac41502ce4b37edadf01556977" exitCode=0 Mar 17 16:09:39 crc kubenswrapper[4767]: I0317 16:09:39.538936 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-c9h7f" event={"ID":"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2","Type":"ContainerDied","Data":"83f91c392163db2a58f38aab200eec75baf8e6ac41502ce4b37edadf01556977"} Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.023696 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.062229 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-combined-ca-bundle\") pod \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.062326 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-scripts\") pod \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.062399 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-config-data\") pod \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.062608 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fprnd\" (UniqueName: \"kubernetes.io/projected/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-kube-api-access-fprnd\") pod \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\" (UID: \"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2\") " Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.095065 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-kube-api-access-fprnd" (OuterVolumeSpecName: "kube-api-access-fprnd") pod "fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2" (UID: "fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2"). InnerVolumeSpecName "kube-api-access-fprnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.108138 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-scripts" (OuterVolumeSpecName: "scripts") pod "fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2" (UID: "fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.111381 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2" (UID: "fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.124709 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-config-data" (OuterVolumeSpecName: "config-data") pod "fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2" (UID: "fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.170039 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.170083 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.170094 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.170103 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fprnd\" (UniqueName: \"kubernetes.io/projected/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2-kube-api-access-fprnd\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.568129 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-c9h7f" event={"ID":"fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2","Type":"ContainerDied","Data":"f5774996fbd96ea283cb2eb306d53e49a0efe0c4eb15da2da4ec44ace667db5c"} Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.568497 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5774996fbd96ea283cb2eb306d53e49a0efe0c4eb15da2da4ec44ace667db5c" Mar 17 16:09:41 crc kubenswrapper[4767]: I0317 16:09:41.568217 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-c9h7f" Mar 17 16:09:42 crc kubenswrapper[4767]: I0317 16:09:42.737458 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Mar 17 16:09:42 crc kubenswrapper[4767]: I0317 16:09:42.738198 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-api" containerID="cri-o://9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7" gracePeriod=30 Mar 17 16:09:42 crc kubenswrapper[4767]: I0317 16:09:42.738890 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-listener" containerID="cri-o://5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775" gracePeriod=30 Mar 17 16:09:42 crc kubenswrapper[4767]: I0317 16:09:42.738959 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-notifier" containerID="cri-o://c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe" gracePeriod=30 Mar 17 16:09:42 crc kubenswrapper[4767]: I0317 16:09:42.739012 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-evaluator" containerID="cri-o://aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c" gracePeriod=30 Mar 17 16:09:43 crc kubenswrapper[4767]: I0317 16:09:43.681131 4767 generic.go:334] "Generic (PLEG): container finished" podID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerID="aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c" exitCode=0 Mar 17 16:09:43 crc kubenswrapper[4767]: I0317 16:09:43.681216 4767 generic.go:334] "Generic (PLEG): container finished" podID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerID="9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7" exitCode=0 Mar 17 16:09:43 crc kubenswrapper[4767]: I0317 16:09:43.681247 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4ec07ded-d23f-411d-a86f-4d971155bf1f","Type":"ContainerDied","Data":"aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c"} Mar 17 16:09:43 crc kubenswrapper[4767]: I0317 16:09:43.681278 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4ec07ded-d23f-411d-a86f-4d971155bf1f","Type":"ContainerDied","Data":"9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7"} Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.696321 4767 generic.go:334] "Generic (PLEG): container finished" podID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" containerID="2f570a2d5d87e91310496d39d3f5743a020b78ef2da12089bcec9dc84e6c63c1" exitCode=0 Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.696574 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"3ad3d357-572b-4b20-bfe6-4aa0b3513d41","Type":"ContainerDied","Data":"2f570a2d5d87e91310496d39d3f5743a020b78ef2da12089bcec9dc84e6c63c1"} Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.696674 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"3ad3d357-572b-4b20-bfe6-4aa0b3513d41","Type":"ContainerDied","Data":"a84ed7d7d41f56cd90679c1b7794d35f967406fcc5a89dac0435df1d24fe42cb"} Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.696690 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a84ed7d7d41f56cd90679c1b7794d35f967406fcc5a89dac0435df1d24fe42cb" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.741682 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.915768 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5dp4\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-kube-api-access-x5dp4\") pod \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.915840 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-erlang-cookie\") pod \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.915912 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-pod-info\") pod \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.916058 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-plugins-conf\") pod \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.916147 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-confd\") pod \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.916901 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3ad3d357-572b-4b20-bfe6-4aa0b3513d41" (UID: "3ad3d357-572b-4b20-bfe6-4aa0b3513d41"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.918425 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\") pod \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.918694 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-plugins\") pod \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.918743 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-tls\") pod \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.918777 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-config-data\") pod \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.918862 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-server-conf\") pod \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.919031 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-erlang-cookie-secret\") pod \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\" (UID: \"3ad3d357-572b-4b20-bfe6-4aa0b3513d41\") " Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.919458 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3ad3d357-572b-4b20-bfe6-4aa0b3513d41" (UID: "3ad3d357-572b-4b20-bfe6-4aa0b3513d41"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.919847 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3ad3d357-572b-4b20-bfe6-4aa0b3513d41" (UID: "3ad3d357-572b-4b20-bfe6-4aa0b3513d41"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.921395 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.921545 4767 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.925396 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.924342 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-pod-info" (OuterVolumeSpecName: "pod-info") pod "3ad3d357-572b-4b20-bfe6-4aa0b3513d41" (UID: "3ad3d357-572b-4b20-bfe6-4aa0b3513d41"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.924855 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "3ad3d357-572b-4b20-bfe6-4aa0b3513d41" (UID: "3ad3d357-572b-4b20-bfe6-4aa0b3513d41"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.926108 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-kube-api-access-x5dp4" (OuterVolumeSpecName: "kube-api-access-x5dp4") pod "3ad3d357-572b-4b20-bfe6-4aa0b3513d41" (UID: "3ad3d357-572b-4b20-bfe6-4aa0b3513d41"). InnerVolumeSpecName "kube-api-access-x5dp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.951411 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3ad3d357-572b-4b20-bfe6-4aa0b3513d41" (UID: "3ad3d357-572b-4b20-bfe6-4aa0b3513d41"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.960791 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3" (OuterVolumeSpecName: "persistence") pod "3ad3d357-572b-4b20-bfe6-4aa0b3513d41" (UID: "3ad3d357-572b-4b20-bfe6-4aa0b3513d41"). InnerVolumeSpecName "pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 16:09:44 crc kubenswrapper[4767]: I0317 16:09:44.985637 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-config-data" (OuterVolumeSpecName: "config-data") pod "3ad3d357-572b-4b20-bfe6-4aa0b3513d41" (UID: "3ad3d357-572b-4b20-bfe6-4aa0b3513d41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.028284 4767 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.028342 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5dp4\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-kube-api-access-x5dp4\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.028356 4767 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-pod-info\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.028403 4767 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\") on node \"crc\" " Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.028422 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.028436 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.050073 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-server-conf" (OuterVolumeSpecName: "server-conf") pod "3ad3d357-572b-4b20-bfe6-4aa0b3513d41" (UID: "3ad3d357-572b-4b20-bfe6-4aa0b3513d41"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.070231 4767 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.070698 4767 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3") on node "crc" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.115269 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3ad3d357-572b-4b20-bfe6-4aa0b3513d41" (UID: "3ad3d357-572b-4b20-bfe6-4aa0b3513d41"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.131705 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.132011 4767 reconciler_common.go:293] "Volume detached for volume \"pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.132099 4767 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3ad3d357-572b-4b20-bfe6-4aa0b3513d41-server-conf\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.714615 4767 generic.go:334] "Generic (PLEG): container finished" podID="bec306cb-266c-4161-86cc-da1f1cb66faa" containerID="49ac648ab09d6b880b8ffd37021fa56228b4fab88173748f63e9206a7b9bded0" exitCode=0 Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.714865 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" event={"ID":"bec306cb-266c-4161-86cc-da1f1cb66faa","Type":"ContainerDied","Data":"49ac648ab09d6b880b8ffd37021fa56228b4fab88173748f63e9206a7b9bded0"} Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.719116 4767 generic.go:334] "Generic (PLEG): container finished" podID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerID="c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe" exitCode=0 Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.719182 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4ec07ded-d23f-411d-a86f-4d971155bf1f","Type":"ContainerDied","Data":"c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe"} Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.719289 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.779403 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.806902 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.834051 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-1"] Mar 17 16:09:45 crc kubenswrapper[4767]: E0317 16:09:45.835240 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" containerName="setup-container" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.835347 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" containerName="setup-container" Mar 17 16:09:45 crc kubenswrapper[4767]: E0317 16:09:45.835491 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2" containerName="aodh-db-sync" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.835573 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2" containerName="aodh-db-sync" Mar 17 16:09:45 crc kubenswrapper[4767]: E0317 16:09:45.835672 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" containerName="rabbitmq" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.835746 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" containerName="rabbitmq" Mar 17 16:09:45 crc kubenswrapper[4767]: E0317 16:09:45.835862 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" containerName="heat-engine" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.835952 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" containerName="heat-engine" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.836454 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2" containerName="aodh-db-sync" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.845509 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" containerName="rabbitmq" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.845679 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c87dcb-97fe-48d4-82b3-fa36fb9c89c6" containerName="heat-engine" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.848075 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 17 16:09:45 crc kubenswrapper[4767]: I0317 16:09:45.867698 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.065012 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a54947c6-19dd-4513-b6be-f72ad14a8bcb-server-conf\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.065505 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a54947c6-19dd-4513-b6be-f72ad14a8bcb-pod-info\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.065529 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.065578 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.065644 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.065690 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.065739 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ctv6\" (UniqueName: \"kubernetes.io/projected/a54947c6-19dd-4513-b6be-f72ad14a8bcb-kube-api-access-7ctv6\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.065767 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54947c6-19dd-4513-b6be-f72ad14a8bcb-config-data\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.065831 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a54947c6-19dd-4513-b6be-f72ad14a8bcb-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.065858 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a54947c6-19dd-4513-b6be-f72ad14a8bcb-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.065980 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.170416 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.170531 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a54947c6-19dd-4513-b6be-f72ad14a8bcb-server-conf\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.170588 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a54947c6-19dd-4513-b6be-f72ad14a8bcb-pod-info\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.170614 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.170645 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.170668 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.170702 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.170741 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ctv6\" (UniqueName: \"kubernetes.io/projected/a54947c6-19dd-4513-b6be-f72ad14a8bcb-kube-api-access-7ctv6\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.170771 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54947c6-19dd-4513-b6be-f72ad14a8bcb-config-data\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.170848 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a54947c6-19dd-4513-b6be-f72ad14a8bcb-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.170876 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a54947c6-19dd-4513-b6be-f72ad14a8bcb-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.173206 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a54947c6-19dd-4513-b6be-f72ad14a8bcb-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.173561 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54947c6-19dd-4513-b6be-f72ad14a8bcb-config-data\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.174401 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a54947c6-19dd-4513-b6be-f72ad14a8bcb-server-conf\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.180575 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.183634 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.184408 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a54947c6-19dd-4513-b6be-f72ad14a8bcb-pod-info\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.185683 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.188953 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a54947c6-19dd-4513-b6be-f72ad14a8bcb-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.191049 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a54947c6-19dd-4513-b6be-f72ad14a8bcb-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.205275 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ctv6\" (UniqueName: \"kubernetes.io/projected/a54947c6-19dd-4513-b6be-f72ad14a8bcb-kube-api-access-7ctv6\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.246372 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.246701 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/084c91f26dfb834887feef1465eec9c6367b6581a3e7afa9bcda1aeaba77086c/globalmount\"" pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.458367 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a45de50-5f38-49d8-aa25-6da880a82ba3\") pod \"rabbitmq-server-1\" (UID: \"a54947c6-19dd-4513-b6be-f72ad14a8bcb\") " pod="openstack/rabbitmq-server-1" Mar 17 16:09:46 crc kubenswrapper[4767]: I0317 16:09:46.650452 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.373310 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ad3d357-572b-4b20-bfe6-4aa0b3513d41" path="/var/lib/kubelet/pods/3ad3d357-572b-4b20-bfe6-4aa0b3513d41/volumes" Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.564055 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.764658 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" event={"ID":"bec306cb-266c-4161-86cc-da1f1cb66faa","Type":"ContainerDied","Data":"87ce1dbca28c3a54aad9dcbc761887a5790c2a4510204b2e1c6a1c340acb6b12"} Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.764725 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87ce1dbca28c3a54aad9dcbc761887a5790c2a4510204b2e1c6a1c340acb6b12" Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.766229 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"a54947c6-19dd-4513-b6be-f72ad14a8bcb","Type":"ContainerStarted","Data":"9d4d8ac22308071e5a5693bdb28ff195efda78a51645880eee005d539c598b43"} Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.844392 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.975893 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjrl7\" (UniqueName: \"kubernetes.io/projected/bec306cb-266c-4161-86cc-da1f1cb66faa-kube-api-access-hjrl7\") pod \"bec306cb-266c-4161-86cc-da1f1cb66faa\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.976138 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-ssh-key-openstack-edpm-ipam\") pod \"bec306cb-266c-4161-86cc-da1f1cb66faa\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.976430 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-inventory\") pod \"bec306cb-266c-4161-86cc-da1f1cb66faa\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.976826 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-repo-setup-combined-ca-bundle\") pod \"bec306cb-266c-4161-86cc-da1f1cb66faa\" (UID: \"bec306cb-266c-4161-86cc-da1f1cb66faa\") " Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.983562 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "bec306cb-266c-4161-86cc-da1f1cb66faa" (UID: "bec306cb-266c-4161-86cc-da1f1cb66faa"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:47 crc kubenswrapper[4767]: I0317 16:09:47.988596 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bec306cb-266c-4161-86cc-da1f1cb66faa-kube-api-access-hjrl7" (OuterVolumeSpecName: "kube-api-access-hjrl7") pod "bec306cb-266c-4161-86cc-da1f1cb66faa" (UID: "bec306cb-266c-4161-86cc-da1f1cb66faa"). InnerVolumeSpecName "kube-api-access-hjrl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.031802 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bec306cb-266c-4161-86cc-da1f1cb66faa" (UID: "bec306cb-266c-4161-86cc-da1f1cb66faa"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.035047 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-inventory" (OuterVolumeSpecName: "inventory") pod "bec306cb-266c-4161-86cc-da1f1cb66faa" (UID: "bec306cb-266c-4161-86cc-da1f1cb66faa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.079316 4767 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.079363 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjrl7\" (UniqueName: \"kubernetes.io/projected/bec306cb-266c-4161-86cc-da1f1cb66faa-kube-api-access-hjrl7\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.079374 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.079385 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bec306cb-266c-4161-86cc-da1f1cb66faa-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.778277 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9pdm5" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.953187 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr"] Mar 17 16:09:48 crc kubenswrapper[4767]: E0317 16:09:48.954184 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bec306cb-266c-4161-86cc-da1f1cb66faa" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.954216 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="bec306cb-266c-4161-86cc-da1f1cb66faa" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.954567 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="bec306cb-266c-4161-86cc-da1f1cb66faa" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.956057 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.959200 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.959827 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.959843 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.960122 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:09:48 crc kubenswrapper[4767]: I0317 16:09:48.969212 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr"] Mar 17 16:09:49 crc kubenswrapper[4767]: I0317 16:09:49.107418 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9qtsr\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:49 crc kubenswrapper[4767]: I0317 16:09:49.107491 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whpfp\" (UniqueName: \"kubernetes.io/projected/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-kube-api-access-whpfp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9qtsr\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:49 crc kubenswrapper[4767]: I0317 16:09:49.108288 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9qtsr\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:49 crc kubenswrapper[4767]: I0317 16:09:49.211434 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9qtsr\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:49 crc kubenswrapper[4767]: I0317 16:09:49.211630 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9qtsr\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:49 crc kubenswrapper[4767]: I0317 16:09:49.211699 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whpfp\" (UniqueName: \"kubernetes.io/projected/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-kube-api-access-whpfp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9qtsr\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:49 crc kubenswrapper[4767]: I0317 16:09:49.218788 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9qtsr\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:49 crc kubenswrapper[4767]: I0317 16:09:49.218905 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9qtsr\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:49 crc kubenswrapper[4767]: I0317 16:09:49.228449 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whpfp\" (UniqueName: \"kubernetes.io/projected/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-kube-api-access-whpfp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9qtsr\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:49 crc kubenswrapper[4767]: I0317 16:09:49.279536 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:49 crc kubenswrapper[4767]: I0317 16:09:49.798357 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"a54947c6-19dd-4513-b6be-f72ad14a8bcb","Type":"ContainerStarted","Data":"8c365f087b6b8db277a8a8fb229312e9508169dc028912b03a83c3c6fb841222"} Mar 17 16:09:50 crc kubenswrapper[4767]: I0317 16:09:50.040323 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr"] Mar 17 16:09:50 crc kubenswrapper[4767]: W0317 16:09:50.044486 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7713aaf_72f2_434c_a4a6_1d4ebd9039d2.slice/crio-0be0f7b4df65a3c2a20d1b96232b7d1c84a8dc428aed19f4b9d55176849f0ad0 WatchSource:0}: Error finding container 0be0f7b4df65a3c2a20d1b96232b7d1c84a8dc428aed19f4b9d55176849f0ad0: Status 404 returned error can't find the container with id 0be0f7b4df65a3c2a20d1b96232b7d1c84a8dc428aed19f4b9d55176849f0ad0 Mar 17 16:09:50 crc kubenswrapper[4767]: I0317 16:09:50.474487 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:09:50 crc kubenswrapper[4767]: I0317 16:09:50.989335 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" event={"ID":"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2","Type":"ContainerStarted","Data":"0be0f7b4df65a3c2a20d1b96232b7d1c84a8dc428aed19f4b9d55176849f0ad0"} Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.788511 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.806661 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnv6j\" (UniqueName: \"kubernetes.io/projected/4ec07ded-d23f-411d-a86f-4d971155bf1f-kube-api-access-mnv6j\") pod \"4ec07ded-d23f-411d-a86f-4d971155bf1f\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.806809 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-public-tls-certs\") pod \"4ec07ded-d23f-411d-a86f-4d971155bf1f\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.806827 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-combined-ca-bundle\") pod \"4ec07ded-d23f-411d-a86f-4d971155bf1f\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.806867 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-config-data\") pod \"4ec07ded-d23f-411d-a86f-4d971155bf1f\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.806996 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-scripts\") pod \"4ec07ded-d23f-411d-a86f-4d971155bf1f\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.807041 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-internal-tls-certs\") pod \"4ec07ded-d23f-411d-a86f-4d971155bf1f\" (UID: \"4ec07ded-d23f-411d-a86f-4d971155bf1f\") " Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.821213 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-scripts" (OuterVolumeSpecName: "scripts") pod "4ec07ded-d23f-411d-a86f-4d971155bf1f" (UID: "4ec07ded-d23f-411d-a86f-4d971155bf1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.893030 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ec07ded-d23f-411d-a86f-4d971155bf1f-kube-api-access-mnv6j" (OuterVolumeSpecName: "kube-api-access-mnv6j") pod "4ec07ded-d23f-411d-a86f-4d971155bf1f" (UID: "4ec07ded-d23f-411d-a86f-4d971155bf1f"). InnerVolumeSpecName "kube-api-access-mnv6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.917721 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnv6j\" (UniqueName: \"kubernetes.io/projected/4ec07ded-d23f-411d-a86f-4d971155bf1f-kube-api-access-mnv6j\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.917762 4767 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.940825 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4ec07ded-d23f-411d-a86f-4d971155bf1f" (UID: "4ec07ded-d23f-411d-a86f-4d971155bf1f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:51 crc kubenswrapper[4767]: I0317 16:09:51.944830 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4ec07ded-d23f-411d-a86f-4d971155bf1f" (UID: "4ec07ded-d23f-411d-a86f-4d971155bf1f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.014739 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" event={"ID":"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2","Type":"ContainerStarted","Data":"acb91ef1a2add63eb8ffcf35b578285f74423a6dbacc8bc290a3ccab1127842d"} Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.027316 4767 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.027364 4767 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.037344 4767 generic.go:334] "Generic (PLEG): container finished" podID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerID="5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775" exitCode=0 Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.037411 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4ec07ded-d23f-411d-a86f-4d971155bf1f","Type":"ContainerDied","Data":"5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775"} Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.037448 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4ec07ded-d23f-411d-a86f-4d971155bf1f","Type":"ContainerDied","Data":"daede6d5b050635880d872508b8da6238cef7b3d1998772b33488a3e45ad907a"} Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.037467 4767 scope.go:117] "RemoveContainer" containerID="5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.037521 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.040519 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" podStartSLOduration=3.617413313 podStartE2EDuration="4.040493307s" podCreationTimestamp="2026-03-17 16:09:48 +0000 UTC" firstStartedPulling="2026-03-17 16:09:50.04729499 +0000 UTC m=+1981.460611037" lastFinishedPulling="2026-03-17 16:09:50.470374984 +0000 UTC m=+1981.883691031" observedRunningTime="2026-03-17 16:09:52.037247895 +0000 UTC m=+1983.450563952" watchObservedRunningTime="2026-03-17 16:09:52.040493307 +0000 UTC m=+1983.453809354" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.082212 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ec07ded-d23f-411d-a86f-4d971155bf1f" (UID: "4ec07ded-d23f-411d-a86f-4d971155bf1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.084940 4767 scope.go:117] "RemoveContainer" containerID="c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.099199 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-config-data" (OuterVolumeSpecName: "config-data") pod "4ec07ded-d23f-411d-a86f-4d971155bf1f" (UID: "4ec07ded-d23f-411d-a86f-4d971155bf1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.113718 4767 scope.go:117] "RemoveContainer" containerID="aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.129699 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.129925 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ec07ded-d23f-411d-a86f-4d971155bf1f-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.151127 4767 scope.go:117] "RemoveContainer" containerID="9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.186369 4767 scope.go:117] "RemoveContainer" containerID="5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775" Mar 17 16:09:52 crc kubenswrapper[4767]: E0317 16:09:52.187160 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775\": container with ID starting with 5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775 not found: ID does not exist" containerID="5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.187248 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775"} err="failed to get container status \"5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775\": rpc error: code = NotFound desc = could not find container \"5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775\": container with ID starting with 5ce85ed909f9dd019cf1dd6f895667523e2a397d497623c67ad58314894fb775 not found: ID does not exist" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.187298 4767 scope.go:117] "RemoveContainer" containerID="c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe" Mar 17 16:09:52 crc kubenswrapper[4767]: E0317 16:09:52.188076 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe\": container with ID starting with c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe not found: ID does not exist" containerID="c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.188133 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe"} err="failed to get container status \"c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe\": rpc error: code = NotFound desc = could not find container \"c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe\": container with ID starting with c787c8f270dfb09c58b5573ec909aace2fb16e19c14b9952b11541def74648fe not found: ID does not exist" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.188165 4767 scope.go:117] "RemoveContainer" containerID="aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c" Mar 17 16:09:52 crc kubenswrapper[4767]: E0317 16:09:52.189082 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c\": container with ID starting with aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c not found: ID does not exist" containerID="aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.189198 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c"} err="failed to get container status \"aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c\": rpc error: code = NotFound desc = could not find container \"aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c\": container with ID starting with aae6ca2b6638106ff907b98f62b98f2e949ba90e495d508eb737f06cfc43851c not found: ID does not exist" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.189298 4767 scope.go:117] "RemoveContainer" containerID="9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7" Mar 17 16:09:52 crc kubenswrapper[4767]: E0317 16:09:52.189862 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7\": container with ID starting with 9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7 not found: ID does not exist" containerID="9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.189887 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7"} err="failed to get container status \"9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7\": rpc error: code = NotFound desc = could not find container \"9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7\": container with ID starting with 9c3564f2e54ddc5289b681276d2ebe6ae8545b20dcf1a287febc91d4cfda7fb7 not found: ID does not exist" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.382222 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.400465 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.421076 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Mar 17 16:09:52 crc kubenswrapper[4767]: E0317 16:09:52.421921 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-listener" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.421941 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-listener" Mar 17 16:09:52 crc kubenswrapper[4767]: E0317 16:09:52.421986 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-api" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.421992 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-api" Mar 17 16:09:52 crc kubenswrapper[4767]: E0317 16:09:52.422025 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-evaluator" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.422032 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-evaluator" Mar 17 16:09:52 crc kubenswrapper[4767]: E0317 16:09:52.422051 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-notifier" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.422056 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-notifier" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.422326 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-listener" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.422351 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-evaluator" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.422360 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-notifier" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.422389 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" containerName="aodh-api" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.425099 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.437770 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.438330 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.438459 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-9wfxq" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.440538 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-config-data\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.440644 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-scripts\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.441149 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt24x\" (UniqueName: \"kubernetes.io/projected/160865a3-d656-4a43-af32-5793ffa3da82-kube-api-access-wt24x\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.441227 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-combined-ca-bundle\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.441282 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-internal-tls-certs\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.441362 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-public-tls-certs\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.444520 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.445427 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.449977 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.544844 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt24x\" (UniqueName: \"kubernetes.io/projected/160865a3-d656-4a43-af32-5793ffa3da82-kube-api-access-wt24x\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.544909 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-combined-ca-bundle\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.544948 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-internal-tls-certs\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.544983 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-public-tls-certs\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.545084 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-config-data\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.545121 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-scripts\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.559674 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-internal-tls-certs\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.559779 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-public-tls-certs\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.560089 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-scripts\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.560106 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-combined-ca-bundle\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.564360 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/160865a3-d656-4a43-af32-5793ffa3da82-config-data\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.570109 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt24x\" (UniqueName: \"kubernetes.io/projected/160865a3-d656-4a43-af32-5793ffa3da82-kube-api-access-wt24x\") pod \"aodh-0\" (UID: \"160865a3-d656-4a43-af32-5793ffa3da82\") " pod="openstack/aodh-0" Mar 17 16:09:52 crc kubenswrapper[4767]: I0317 16:09:52.753810 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 17 16:09:53 crc kubenswrapper[4767]: I0317 16:09:53.313208 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 17 16:09:53 crc kubenswrapper[4767]: I0317 16:09:53.370656 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ec07ded-d23f-411d-a86f-4d971155bf1f" path="/var/lib/kubelet/pods/4ec07ded-d23f-411d-a86f-4d971155bf1f/volumes" Mar 17 16:09:54 crc kubenswrapper[4767]: I0317 16:09:54.078974 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"160865a3-d656-4a43-af32-5793ffa3da82","Type":"ContainerStarted","Data":"e2bc4f974e90df2caa1bc756e3eaf508d993fe609d35fd0e62de78c784502431"} Mar 17 16:09:54 crc kubenswrapper[4767]: I0317 16:09:54.079620 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"160865a3-d656-4a43-af32-5793ffa3da82","Type":"ContainerStarted","Data":"115aeec3001a5691b5225c4dec9cebd8c930ee36b87172cd198781d5ca24b932"} Mar 17 16:09:55 crc kubenswrapper[4767]: I0317 16:09:55.178621 4767 generic.go:334] "Generic (PLEG): container finished" podID="c7713aaf-72f2-434c-a4a6-1d4ebd9039d2" containerID="acb91ef1a2add63eb8ffcf35b578285f74423a6dbacc8bc290a3ccab1127842d" exitCode=0 Mar 17 16:09:55 crc kubenswrapper[4767]: I0317 16:09:55.178875 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" event={"ID":"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2","Type":"ContainerDied","Data":"acb91ef1a2add63eb8ffcf35b578285f74423a6dbacc8bc290a3ccab1127842d"} Mar 17 16:09:56 crc kubenswrapper[4767]: I0317 16:09:56.204410 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"160865a3-d656-4a43-af32-5793ffa3da82","Type":"ContainerStarted","Data":"8aac98777fd8795b248d28030cea7c0bbf0ecf96dd0c13324435fcb508c379b6"} Mar 17 16:09:56 crc kubenswrapper[4767]: I0317 16:09:56.921623 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.091015 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-ssh-key-openstack-edpm-ipam\") pod \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.091779 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-inventory\") pod \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.092021 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whpfp\" (UniqueName: \"kubernetes.io/projected/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-kube-api-access-whpfp\") pod \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\" (UID: \"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2\") " Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.101615 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-kube-api-access-whpfp" (OuterVolumeSpecName: "kube-api-access-whpfp") pod "c7713aaf-72f2-434c-a4a6-1d4ebd9039d2" (UID: "c7713aaf-72f2-434c-a4a6-1d4ebd9039d2"). InnerVolumeSpecName "kube-api-access-whpfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.136784 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-inventory" (OuterVolumeSpecName: "inventory") pod "c7713aaf-72f2-434c-a4a6-1d4ebd9039d2" (UID: "c7713aaf-72f2-434c-a4a6-1d4ebd9039d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.148131 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c7713aaf-72f2-434c-a4a6-1d4ebd9039d2" (UID: "c7713aaf-72f2-434c-a4a6-1d4ebd9039d2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.198717 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.198828 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.198890 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whpfp\" (UniqueName: \"kubernetes.io/projected/c7713aaf-72f2-434c-a4a6-1d4ebd9039d2-kube-api-access-whpfp\") on node \"crc\" DevicePath \"\"" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.228017 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"160865a3-d656-4a43-af32-5793ffa3da82","Type":"ContainerStarted","Data":"027d3acf0d96fe65584ed7a7d58c5e2c33815278c3d0f2a408aef2f0f5a87069"} Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.245758 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" event={"ID":"c7713aaf-72f2-434c-a4a6-1d4ebd9039d2","Type":"ContainerDied","Data":"0be0f7b4df65a3c2a20d1b96232b7d1c84a8dc428aed19f4b9d55176849f0ad0"} Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.246161 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0be0f7b4df65a3c2a20d1b96232b7d1c84a8dc428aed19f4b9d55176849f0ad0" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.245820 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9qtsr" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.307863 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt"] Mar 17 16:09:57 crc kubenswrapper[4767]: E0317 16:09:57.308860 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7713aaf-72f2-434c-a4a6-1d4ebd9039d2" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.308887 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7713aaf-72f2-434c-a4a6-1d4ebd9039d2" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.309335 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7713aaf-72f2-434c-a4a6-1d4ebd9039d2" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.310762 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.314967 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.317530 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.317772 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.317785 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.374476 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt"] Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.509322 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.509706 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.509946 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb5pp\" (UniqueName: \"kubernetes.io/projected/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-kube-api-access-bb5pp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.510258 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.613538 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb5pp\" (UniqueName: \"kubernetes.io/projected/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-kube-api-access-bb5pp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.613652 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.613825 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.613898 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.627517 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.631032 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.639470 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.666016 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb5pp\" (UniqueName: \"kubernetes.io/projected/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-kube-api-access-bb5pp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:57 crc kubenswrapper[4767]: I0317 16:09:57.943662 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:09:58 crc kubenswrapper[4767]: I0317 16:09:58.851316 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt"] Mar 17 16:09:59 crc kubenswrapper[4767]: I0317 16:09:59.308467 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" event={"ID":"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd","Type":"ContainerStarted","Data":"106e42f05f5c4d82538bcac32b72638009b70e4c13701852095714ea1d7352cf"} Mar 17 16:09:59 crc kubenswrapper[4767]: I0317 16:09:59.312892 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"160865a3-d656-4a43-af32-5793ffa3da82","Type":"ContainerStarted","Data":"ea1fef60dba7400e95ac69101289b82512b712762292a4809aa076021b3f97b1"} Mar 17 16:09:59 crc kubenswrapper[4767]: I0317 16:09:59.339860 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.546955444 podStartE2EDuration="7.33983234s" podCreationTimestamp="2026-03-17 16:09:52 +0000 UTC" firstStartedPulling="2026-03-17 16:09:53.319251162 +0000 UTC m=+1984.732567209" lastFinishedPulling="2026-03-17 16:09:58.112128058 +0000 UTC m=+1989.525444105" observedRunningTime="2026-03-17 16:09:59.334475588 +0000 UTC m=+1990.747791655" watchObservedRunningTime="2026-03-17 16:09:59.33983234 +0000 UTC m=+1990.753148387" Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.158669 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562730-nw2dp"] Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.161117 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562730-nw2dp" Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.163802 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.166361 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.166433 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.187289 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562730-nw2dp"] Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.214467 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m8b8\" (UniqueName: \"kubernetes.io/projected/49eb134e-cc05-459a-959b-e37e432d642a-kube-api-access-9m8b8\") pod \"auto-csr-approver-29562730-nw2dp\" (UID: \"49eb134e-cc05-459a-959b-e37e432d642a\") " pod="openshift-infra/auto-csr-approver-29562730-nw2dp" Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.317360 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m8b8\" (UniqueName: \"kubernetes.io/projected/49eb134e-cc05-459a-959b-e37e432d642a-kube-api-access-9m8b8\") pod \"auto-csr-approver-29562730-nw2dp\" (UID: \"49eb134e-cc05-459a-959b-e37e432d642a\") " pod="openshift-infra/auto-csr-approver-29562730-nw2dp" Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.330162 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" event={"ID":"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd","Type":"ContainerStarted","Data":"0066ca8ea83821688b5a754dce3b98eec35e082d143fdfda8902694395ff6c89"} Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.343109 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m8b8\" (UniqueName: \"kubernetes.io/projected/49eb134e-cc05-459a-959b-e37e432d642a-kube-api-access-9m8b8\") pod \"auto-csr-approver-29562730-nw2dp\" (UID: \"49eb134e-cc05-459a-959b-e37e432d642a\") " pod="openshift-infra/auto-csr-approver-29562730-nw2dp" Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.361304 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" podStartSLOduration=2.7493683410000003 podStartE2EDuration="3.361275636s" podCreationTimestamp="2026-03-17 16:09:57 +0000 UTC" firstStartedPulling="2026-03-17 16:09:58.847036203 +0000 UTC m=+1990.260352240" lastFinishedPulling="2026-03-17 16:09:59.458943498 +0000 UTC m=+1990.872259535" observedRunningTime="2026-03-17 16:10:00.347203396 +0000 UTC m=+1991.760519463" watchObservedRunningTime="2026-03-17 16:10:00.361275636 +0000 UTC m=+1991.774591673" Mar 17 16:10:00 crc kubenswrapper[4767]: I0317 16:10:00.487416 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562730-nw2dp" Mar 17 16:10:01 crc kubenswrapper[4767]: I0317 16:10:01.154335 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562730-nw2dp"] Mar 17 16:10:01 crc kubenswrapper[4767]: I0317 16:10:01.346214 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562730-nw2dp" event={"ID":"49eb134e-cc05-459a-959b-e37e432d642a","Type":"ContainerStarted","Data":"de1cba95934111051fdfe1f73bade75dcfcbf41b8c67bfa4802f417ffb194803"} Mar 17 16:10:03 crc kubenswrapper[4767]: I0317 16:10:03.382896 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562730-nw2dp" event={"ID":"49eb134e-cc05-459a-959b-e37e432d642a","Type":"ContainerStarted","Data":"f02b1b789c7fab91bfdcb241feccd8e4d944cf6ae3e4f0e3064689dcd7c3742b"} Mar 17 16:10:03 crc kubenswrapper[4767]: I0317 16:10:03.414712 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562730-nw2dp" podStartSLOduration=1.9592098999999998 podStartE2EDuration="3.414684682s" podCreationTimestamp="2026-03-17 16:10:00 +0000 UTC" firstStartedPulling="2026-03-17 16:10:01.131059143 +0000 UTC m=+1992.544375190" lastFinishedPulling="2026-03-17 16:10:02.586533925 +0000 UTC m=+1993.999849972" observedRunningTime="2026-03-17 16:10:03.406232561 +0000 UTC m=+1994.819548618" watchObservedRunningTime="2026-03-17 16:10:03.414684682 +0000 UTC m=+1994.828000749" Mar 17 16:10:05 crc kubenswrapper[4767]: I0317 16:10:05.415226 4767 generic.go:334] "Generic (PLEG): container finished" podID="49eb134e-cc05-459a-959b-e37e432d642a" containerID="f02b1b789c7fab91bfdcb241feccd8e4d944cf6ae3e4f0e3064689dcd7c3742b" exitCode=0 Mar 17 16:10:05 crc kubenswrapper[4767]: I0317 16:10:05.415422 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562730-nw2dp" event={"ID":"49eb134e-cc05-459a-959b-e37e432d642a","Type":"ContainerDied","Data":"f02b1b789c7fab91bfdcb241feccd8e4d944cf6ae3e4f0e3064689dcd7c3742b"} Mar 17 16:10:06 crc kubenswrapper[4767]: I0317 16:10:06.946806 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562730-nw2dp" Mar 17 16:10:07 crc kubenswrapper[4767]: I0317 16:10:07.126090 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m8b8\" (UniqueName: \"kubernetes.io/projected/49eb134e-cc05-459a-959b-e37e432d642a-kube-api-access-9m8b8\") pod \"49eb134e-cc05-459a-959b-e37e432d642a\" (UID: \"49eb134e-cc05-459a-959b-e37e432d642a\") " Mar 17 16:10:07 crc kubenswrapper[4767]: I0317 16:10:07.136932 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49eb134e-cc05-459a-959b-e37e432d642a-kube-api-access-9m8b8" (OuterVolumeSpecName: "kube-api-access-9m8b8") pod "49eb134e-cc05-459a-959b-e37e432d642a" (UID: "49eb134e-cc05-459a-959b-e37e432d642a"). InnerVolumeSpecName "kube-api-access-9m8b8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:10:07 crc kubenswrapper[4767]: I0317 16:10:07.231255 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m8b8\" (UniqueName: \"kubernetes.io/projected/49eb134e-cc05-459a-959b-e37e432d642a-kube-api-access-9m8b8\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:07 crc kubenswrapper[4767]: I0317 16:10:07.448012 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562730-nw2dp" event={"ID":"49eb134e-cc05-459a-959b-e37e432d642a","Type":"ContainerDied","Data":"de1cba95934111051fdfe1f73bade75dcfcbf41b8c67bfa4802f417ffb194803"} Mar 17 16:10:07 crc kubenswrapper[4767]: I0317 16:10:07.448075 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de1cba95934111051fdfe1f73bade75dcfcbf41b8c67bfa4802f417ffb194803" Mar 17 16:10:07 crc kubenswrapper[4767]: I0317 16:10:07.448152 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562730-nw2dp" Mar 17 16:10:07 crc kubenswrapper[4767]: I0317 16:10:07.512356 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562724-wqk5x"] Mar 17 16:10:07 crc kubenswrapper[4767]: I0317 16:10:07.528325 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562724-wqk5x"] Mar 17 16:10:09 crc kubenswrapper[4767]: I0317 16:10:09.372966 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b1f5965-725f-47c1-bfc3-cdb8a3b7a696" path="/var/lib/kubelet/pods/4b1f5965-725f-47c1-bfc3-cdb8a3b7a696/volumes" Mar 17 16:10:15 crc kubenswrapper[4767]: I0317 16:10:15.923663 4767 scope.go:117] "RemoveContainer" containerID="b1a0f869b4fcd37a4835c9b4c678d18fce3210feb179b919f81eef27e69fce14" Mar 17 16:10:15 crc kubenswrapper[4767]: I0317 16:10:15.983532 4767 scope.go:117] "RemoveContainer" containerID="d66bf9f482f6948bd0dcab70f3bcd2ceecc90f9ef84e29dc11705235ee9f82d5" Mar 17 16:10:16 crc kubenswrapper[4767]: I0317 16:10:16.204292 4767 scope.go:117] "RemoveContainer" containerID="f2a60414104af01ed05aeb1eb576cad1646f14e32efe4085e03d6c5de1aefcdb" Mar 17 16:10:16 crc kubenswrapper[4767]: I0317 16:10:16.253817 4767 scope.go:117] "RemoveContainer" containerID="d578683d641746e5a92c34391fd4826e7a2330cdbaf5dabe1c682894cf1d52c6" Mar 17 16:10:16 crc kubenswrapper[4767]: I0317 16:10:16.305835 4767 scope.go:117] "RemoveContainer" containerID="bfb2e2fa71eea606d992e9bc0235bcb0151201ab641f392b4e91ac2749bdc185" Mar 17 16:10:16 crc kubenswrapper[4767]: I0317 16:10:16.387371 4767 scope.go:117] "RemoveContainer" containerID="2f570a2d5d87e91310496d39d3f5743a020b78ef2da12089bcec9dc84e6c63c1" Mar 17 16:10:16 crc kubenswrapper[4767]: I0317 16:10:16.447762 4767 scope.go:117] "RemoveContainer" containerID="1316e141dd3fbf0e6f1aa7a85bea7b42f5c42149752fae5a0298bb94b6124770" Mar 17 16:10:16 crc kubenswrapper[4767]: I0317 16:10:16.475926 4767 scope.go:117] "RemoveContainer" containerID="fb68641df086d68c697275b20dc2997ea82074be958cf89d4c1031793e6d4fcb" Mar 17 16:10:22 crc kubenswrapper[4767]: I0317 16:10:22.704537 4767 generic.go:334] "Generic (PLEG): container finished" podID="a54947c6-19dd-4513-b6be-f72ad14a8bcb" containerID="8c365f087b6b8db277a8a8fb229312e9508169dc028912b03a83c3c6fb841222" exitCode=0 Mar 17 16:10:22 crc kubenswrapper[4767]: I0317 16:10:22.705087 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"a54947c6-19dd-4513-b6be-f72ad14a8bcb","Type":"ContainerDied","Data":"8c365f087b6b8db277a8a8fb229312e9508169dc028912b03a83c3c6fb841222"} Mar 17 16:10:23 crc kubenswrapper[4767]: I0317 16:10:23.722690 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"a54947c6-19dd-4513-b6be-f72ad14a8bcb","Type":"ContainerStarted","Data":"236707390d9c40a7b1e817ccd4165f28bcacedb37213fe3067913ce16fa8f228"} Mar 17 16:10:23 crc kubenswrapper[4767]: I0317 16:10:23.723686 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-1" Mar 17 16:10:23 crc kubenswrapper[4767]: I0317 16:10:23.778849 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-1" podStartSLOduration=38.778813663 podStartE2EDuration="38.778813663s" podCreationTimestamp="2026-03-17 16:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:10:23.74880435 +0000 UTC m=+2015.162120407" watchObservedRunningTime="2026-03-17 16:10:23.778813663 +0000 UTC m=+2015.192129710" Mar 17 16:10:36 crc kubenswrapper[4767]: I0317 16:10:36.654473 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-1" Mar 17 16:10:36 crc kubenswrapper[4767]: I0317 16:10:36.713763 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 16:10:41 crc kubenswrapper[4767]: I0317 16:10:41.930351 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="048fa918-95b2-404e-81f5-8989ed47ec0a" containerName="rabbitmq" containerID="cri-o://b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74" gracePeriod=604795 Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.177732 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.246372 4767 generic.go:334] "Generic (PLEG): container finished" podID="048fa918-95b2-404e-81f5-8989ed47ec0a" containerID="b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74" exitCode=0 Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.246439 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"048fa918-95b2-404e-81f5-8989ed47ec0a","Type":"ContainerDied","Data":"b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74"} Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.246485 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"048fa918-95b2-404e-81f5-8989ed47ec0a","Type":"ContainerDied","Data":"4ff0b4a1c146a1bbe72ca66e23a27c07c1b21b664e2f47a9b254a15aa26541d4"} Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.246515 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.246541 4767 scope.go:117] "RemoveContainer" containerID="b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.350122 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-server-conf\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.350514 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-tls\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.350591 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-erlang-cookie\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.350658 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-plugins\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.350741 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-config-data\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.351474 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.351530 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/048fa918-95b2-404e-81f5-8989ed47ec0a-erlang-cookie-secret\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.351631 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-plugins-conf\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.351675 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/048fa918-95b2-404e-81f5-8989ed47ec0a-pod-info\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.351817 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-confd\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.351880 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnqh9\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-kube-api-access-rnqh9\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.352470 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.353012 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.353246 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.353263 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.355602 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.365546 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/048fa918-95b2-404e-81f5-8989ed47ec0a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.381251 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.412138 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/048fa918-95b2-404e-81f5-8989ed47ec0a-pod-info" (OuterVolumeSpecName: "pod-info") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.427871 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-kube-api-access-rnqh9" (OuterVolumeSpecName: "kube-api-access-rnqh9") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a"). InnerVolumeSpecName "kube-api-access-rnqh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.450360 4767 scope.go:117] "RemoveContainer" containerID="90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.466894 4767 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/048fa918-95b2-404e-81f5-8989ed47ec0a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.466926 4767 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.466958 4767 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/048fa918-95b2-404e-81f5-8989ed47ec0a-pod-info\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.466969 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnqh9\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-kube-api-access-rnqh9\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.466980 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.489819 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-config-data" (OuterVolumeSpecName: "config-data") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:10:49 crc kubenswrapper[4767]: E0317 16:10:49.527692 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0 podName:048fa918-95b2-404e-81f5-8989ed47ec0a nodeName:}" failed. No retries permitted until 2026-03-17 16:10:50.027655443 +0000 UTC m=+2041.440971490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "persistence" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a") : kubernetes.io/csi: Unmounter.TearDownAt failed: rpc error: code = Unknown desc = check target path: could not get consistent content of /proc/mounts after 3 attempts Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.537637 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-server-conf" (OuterVolumeSpecName: "server-conf") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.570757 4767 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-server-conf\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.571138 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/048fa918-95b2-404e-81f5-8989ed47ec0a-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.656120 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.675041 4767 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/048fa918-95b2-404e-81f5-8989ed47ec0a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.767049 4767 scope.go:117] "RemoveContainer" containerID="b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74" Mar 17 16:10:49 crc kubenswrapper[4767]: E0317 16:10:49.767686 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74\": container with ID starting with b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74 not found: ID does not exist" containerID="b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.767744 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74"} err="failed to get container status \"b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74\": rpc error: code = NotFound desc = could not find container \"b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74\": container with ID starting with b9cbcb347a1945bff68c48488cb012e19f2ac5e22cfa52eaca180d7b2396df74 not found: ID does not exist" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.767795 4767 scope.go:117] "RemoveContainer" containerID="90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9" Mar 17 16:10:49 crc kubenswrapper[4767]: E0317 16:10:49.768277 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9\": container with ID starting with 90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9 not found: ID does not exist" containerID="90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9" Mar 17 16:10:49 crc kubenswrapper[4767]: I0317 16:10:49.768301 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9"} err="failed to get container status \"90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9\": rpc error: code = NotFound desc = could not find container \"90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9\": container with ID starting with 90d7b776269b9dda06c61f6d728a5f33f0b97977d5b5fbdc13e02857befcbcd9 not found: ID does not exist" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.087040 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") pod \"048fa918-95b2-404e-81f5-8989ed47ec0a\" (UID: \"048fa918-95b2-404e-81f5-8989ed47ec0a\") " Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.121124 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0" (OuterVolumeSpecName: "persistence") pod "048fa918-95b2-404e-81f5-8989ed47ec0a" (UID: "048fa918-95b2-404e-81f5-8989ed47ec0a"). InnerVolumeSpecName "pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.192082 4767 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") on node \"crc\" " Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.195903 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.216599 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.235625 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 16:10:50 crc kubenswrapper[4767]: E0317 16:10:50.238469 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49eb134e-cc05-459a-959b-e37e432d642a" containerName="oc" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.238498 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="49eb134e-cc05-459a-959b-e37e432d642a" containerName="oc" Mar 17 16:10:50 crc kubenswrapper[4767]: E0317 16:10:50.238528 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048fa918-95b2-404e-81f5-8989ed47ec0a" containerName="setup-container" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.238536 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="048fa918-95b2-404e-81f5-8989ed47ec0a" containerName="setup-container" Mar 17 16:10:50 crc kubenswrapper[4767]: E0317 16:10:50.238548 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048fa918-95b2-404e-81f5-8989ed47ec0a" containerName="rabbitmq" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.238556 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="048fa918-95b2-404e-81f5-8989ed47ec0a" containerName="rabbitmq" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.238847 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="048fa918-95b2-404e-81f5-8989ed47ec0a" containerName="rabbitmq" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.238876 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="49eb134e-cc05-459a-959b-e37e432d642a" containerName="oc" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.239894 4767 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.240123 4767 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0") on node "crc" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.240762 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.283699 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.297315 4767 reconciler_common.go:293] "Volume detached for volume \"pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") on node \"crc\" DevicePath \"\"" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.399668 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glgwk\" (UniqueName: \"kubernetes.io/projected/f4d81837-af8b-49f9-8568-c8bef5c3773e-kube-api-access-glgwk\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.399735 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.399869 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.399921 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.399964 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4d81837-af8b-49f9-8568-c8bef5c3773e-config-data\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.400119 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.400199 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.400267 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f4d81837-af8b-49f9-8568-c8bef5c3773e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.400402 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f4d81837-af8b-49f9-8568-c8bef5c3773e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.400469 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f4d81837-af8b-49f9-8568-c8bef5c3773e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.400786 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f4d81837-af8b-49f9-8568-c8bef5c3773e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.504308 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glgwk\" (UniqueName: \"kubernetes.io/projected/f4d81837-af8b-49f9-8568-c8bef5c3773e-kube-api-access-glgwk\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.504370 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.504421 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.504451 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.504487 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4d81837-af8b-49f9-8568-c8bef5c3773e-config-data\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.504587 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.504672 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.504699 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f4d81837-af8b-49f9-8568-c8bef5c3773e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.504744 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f4d81837-af8b-49f9-8568-c8bef5c3773e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.504774 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f4d81837-af8b-49f9-8568-c8bef5c3773e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.504879 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f4d81837-af8b-49f9-8568-c8bef5c3773e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.505000 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.505463 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.506450 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4d81837-af8b-49f9-8568-c8bef5c3773e-config-data\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.506702 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f4d81837-af8b-49f9-8568-c8bef5c3773e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.507127 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f4d81837-af8b-49f9-8568-c8bef5c3773e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.507589 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.507622 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8bdc0cdd497bb3d78a3d0a8594963d237808ef493d9599e21e04f965565170a3/globalmount\"" pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.509868 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f4d81837-af8b-49f9-8568-c8bef5c3773e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.511222 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.511366 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f4d81837-af8b-49f9-8568-c8bef5c3773e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.517247 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f4d81837-af8b-49f9-8568-c8bef5c3773e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.523982 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glgwk\" (UniqueName: \"kubernetes.io/projected/f4d81837-af8b-49f9-8568-c8bef5c3773e-kube-api-access-glgwk\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.597015 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-caa660e3-48aa-49df-8ed1-0bc9bf9cbfe0\") pod \"rabbitmq-server-0\" (UID: \"f4d81837-af8b-49f9-8568-c8bef5c3773e\") " pod="openstack/rabbitmq-server-0" Mar 17 16:10:50 crc kubenswrapper[4767]: I0317 16:10:50.883266 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 16:10:51 crc kubenswrapper[4767]: I0317 16:10:51.370724 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="048fa918-95b2-404e-81f5-8989ed47ec0a" path="/var/lib/kubelet/pods/048fa918-95b2-404e-81f5-8989ed47ec0a/volumes" Mar 17 16:10:51 crc kubenswrapper[4767]: I0317 16:10:51.439917 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 16:10:52 crc kubenswrapper[4767]: I0317 16:10:52.308613 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f4d81837-af8b-49f9-8568-c8bef5c3773e","Type":"ContainerStarted","Data":"57ae5af01d4e44e232b19f3d96981d161f182fe55a4a902a988125ebcda6b86c"} Mar 17 16:10:54 crc kubenswrapper[4767]: I0317 16:10:54.410039 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f4d81837-af8b-49f9-8568-c8bef5c3773e","Type":"ContainerStarted","Data":"1523914b84e8fba838452b74839f775f371f3ed122eb105ba2a7dd929f5c3865"} Mar 17 16:11:16 crc kubenswrapper[4767]: I0317 16:11:16.713113 4767 scope.go:117] "RemoveContainer" containerID="c7d4dd32069ee1072975d0601b1ffa8e62d738ddbb5023ad4b3f2b088b3efcdf" Mar 17 16:11:17 crc kubenswrapper[4767]: I0317 16:11:17.053503 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vhvs8"] Mar 17 16:11:17 crc kubenswrapper[4767]: I0317 16:11:17.071620 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vhvs8"] Mar 17 16:11:17 crc kubenswrapper[4767]: I0317 16:11:17.370446 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82b61927-c834-4a58-a140-c2c63686f6f1" path="/var/lib/kubelet/pods/82b61927-c834-4a58-a140-c2c63686f6f1/volumes" Mar 17 16:11:20 crc kubenswrapper[4767]: I0317 16:11:20.039233 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-ccf0-account-create-update-vzhjl"] Mar 17 16:11:20 crc kubenswrapper[4767]: I0317 16:11:20.054076 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-ccf0-account-create-update-vzhjl"] Mar 17 16:11:21 crc kubenswrapper[4767]: I0317 16:11:21.041509 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-c428-account-create-update-x48t4"] Mar 17 16:11:21 crc kubenswrapper[4767]: I0317 16:11:21.063712 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-c428-account-create-update-x48t4"] Mar 17 16:11:21 crc kubenswrapper[4767]: I0317 16:11:21.309701 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9758-account-create-update-xjjlg"] Mar 17 16:11:21 crc kubenswrapper[4767]: I0317 16:11:21.328071 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-qwb25"] Mar 17 16:11:21 crc kubenswrapper[4767]: I0317 16:11:21.341098 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9758-account-create-update-xjjlg"] Mar 17 16:11:21 crc kubenswrapper[4767]: I0317 16:11:21.350926 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-qwb25"] Mar 17 16:11:21 crc kubenswrapper[4767]: I0317 16:11:21.401735 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22e12d8e-e55f-4839-afdb-a03c60523aed" path="/var/lib/kubelet/pods/22e12d8e-e55f-4839-afdb-a03c60523aed/volumes" Mar 17 16:11:21 crc kubenswrapper[4767]: I0317 16:11:21.405565 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf" path="/var/lib/kubelet/pods/85ad80cc-d138-4f6e-91d0-fc5cf56e5bdf/volumes" Mar 17 16:11:21 crc kubenswrapper[4767]: I0317 16:11:21.410705 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4632708-aa04-4b74-bdda-99650680eac8" path="/var/lib/kubelet/pods/d4632708-aa04-4b74-bdda-99650680eac8/volumes" Mar 17 16:11:21 crc kubenswrapper[4767]: I0317 16:11:21.413839 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f70029ee-d0ed-4aaa-8907-f3bfb4b196b3" path="/var/lib/kubelet/pods/f70029ee-d0ed-4aaa-8907-f3bfb4b196b3/volumes" Mar 17 16:11:26 crc kubenswrapper[4767]: I0317 16:11:26.043315 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-3aef-account-create-update-x9gj7"] Mar 17 16:11:26 crc kubenswrapper[4767]: I0317 16:11:26.059120 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-kplsg"] Mar 17 16:11:26 crc kubenswrapper[4767]: I0317 16:11:26.081887 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-n9ftr"] Mar 17 16:11:26 crc kubenswrapper[4767]: I0317 16:11:26.094118 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-n9ftr"] Mar 17 16:11:26 crc kubenswrapper[4767]: I0317 16:11:26.111733 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-3aef-account-create-update-x9gj7"] Mar 17 16:11:26 crc kubenswrapper[4767]: I0317 16:11:26.125773 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-kplsg"] Mar 17 16:11:27 crc kubenswrapper[4767]: I0317 16:11:27.368877 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33ed389a-fa13-445f-826e-a88c04ef059d" path="/var/lib/kubelet/pods/33ed389a-fa13-445f-826e-a88c04ef059d/volumes" Mar 17 16:11:27 crc kubenswrapper[4767]: I0317 16:11:27.371008 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae4f3fba-7231-451f-a3cd-d560c85212dd" path="/var/lib/kubelet/pods/ae4f3fba-7231-451f-a3cd-d560c85212dd/volumes" Mar 17 16:11:27 crc kubenswrapper[4767]: I0317 16:11:27.373253 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f745e6a5-fa5d-4bee-9df5-df119632fbf0" path="/var/lib/kubelet/pods/f745e6a5-fa5d-4bee-9df5-df119632fbf0/volumes" Mar 17 16:11:27 crc kubenswrapper[4767]: I0317 16:11:27.438052 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4d81837-af8b-49f9-8568-c8bef5c3773e" containerID="1523914b84e8fba838452b74839f775f371f3ed122eb105ba2a7dd929f5c3865" exitCode=0 Mar 17 16:11:27 crc kubenswrapper[4767]: I0317 16:11:27.438108 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f4d81837-af8b-49f9-8568-c8bef5c3773e","Type":"ContainerDied","Data":"1523914b84e8fba838452b74839f775f371f3ed122eb105ba2a7dd929f5c3865"} Mar 17 16:11:28 crc kubenswrapper[4767]: I0317 16:11:28.459836 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f4d81837-af8b-49f9-8568-c8bef5c3773e","Type":"ContainerStarted","Data":"020d66a67a0c832648320701e305551b3bcb1d625de768e519ea9cb330ddddf4"} Mar 17 16:11:28 crc kubenswrapper[4767]: I0317 16:11:28.460852 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 17 16:11:28 crc kubenswrapper[4767]: I0317 16:11:28.501750 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.501722226 podStartE2EDuration="38.501722226s" podCreationTimestamp="2026-03-17 16:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:11:28.48991138 +0000 UTC m=+2079.903227437" watchObservedRunningTime="2026-03-17 16:11:28.501722226 +0000 UTC m=+2079.915038283" Mar 17 16:11:29 crc kubenswrapper[4767]: I0317 16:11:29.044878 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-fa1b-account-create-update-nkxnj"] Mar 17 16:11:29 crc kubenswrapper[4767]: I0317 16:11:29.066892 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-fa1b-account-create-update-nkxnj"] Mar 17 16:11:29 crc kubenswrapper[4767]: I0317 16:11:29.523529 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89f09b66-7eec-4ad8-9116-83fa9d80fead" path="/var/lib/kubelet/pods/89f09b66-7eec-4ad8-9116-83fa9d80fead/volumes" Mar 17 16:11:30 crc kubenswrapper[4767]: I0317 16:11:30.036013 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f"] Mar 17 16:11:30 crc kubenswrapper[4767]: I0317 16:11:30.049143 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-w9n2f"] Mar 17 16:11:31 crc kubenswrapper[4767]: I0317 16:11:31.370947 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e718a8e4-5847-4051-a7d3-d01f0c22d801" path="/var/lib/kubelet/pods/e718a8e4-5847-4051-a7d3-d01f0c22d801/volumes" Mar 17 16:11:33 crc kubenswrapper[4767]: I0317 16:11:33.055221 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-xt87s"] Mar 17 16:11:33 crc kubenswrapper[4767]: I0317 16:11:33.066105 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-xt87s"] Mar 17 16:11:33 crc kubenswrapper[4767]: I0317 16:11:33.380113 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff7518ae-245d-419e-9328-58420b43d6a7" path="/var/lib/kubelet/pods/ff7518ae-245d-419e-9328-58420b43d6a7/volumes" Mar 17 16:11:34 crc kubenswrapper[4767]: I0317 16:11:34.166597 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:11:34 crc kubenswrapper[4767]: I0317 16:11:34.167017 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:11:40 crc kubenswrapper[4767]: I0317 16:11:40.888356 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 17 16:12:00 crc kubenswrapper[4767]: I0317 16:12:00.175710 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562732-s2frn"] Mar 17 16:12:00 crc kubenswrapper[4767]: I0317 16:12:00.179003 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562732-s2frn" Mar 17 16:12:00 crc kubenswrapper[4767]: I0317 16:12:00.183773 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:12:00 crc kubenswrapper[4767]: I0317 16:12:00.184001 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:12:00 crc kubenswrapper[4767]: I0317 16:12:00.184219 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:12:00 crc kubenswrapper[4767]: I0317 16:12:00.191089 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562732-s2frn"] Mar 17 16:12:00 crc kubenswrapper[4767]: I0317 16:12:00.310893 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxhqv\" (UniqueName: \"kubernetes.io/projected/1af290f7-a893-41d9-87ce-4e6f42d73e28-kube-api-access-bxhqv\") pod \"auto-csr-approver-29562732-s2frn\" (UID: \"1af290f7-a893-41d9-87ce-4e6f42d73e28\") " pod="openshift-infra/auto-csr-approver-29562732-s2frn" Mar 17 16:12:00 crc kubenswrapper[4767]: I0317 16:12:00.413762 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxhqv\" (UniqueName: \"kubernetes.io/projected/1af290f7-a893-41d9-87ce-4e6f42d73e28-kube-api-access-bxhqv\") pod \"auto-csr-approver-29562732-s2frn\" (UID: \"1af290f7-a893-41d9-87ce-4e6f42d73e28\") " pod="openshift-infra/auto-csr-approver-29562732-s2frn" Mar 17 16:12:00 crc kubenswrapper[4767]: I0317 16:12:00.441305 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxhqv\" (UniqueName: \"kubernetes.io/projected/1af290f7-a893-41d9-87ce-4e6f42d73e28-kube-api-access-bxhqv\") pod \"auto-csr-approver-29562732-s2frn\" (UID: \"1af290f7-a893-41d9-87ce-4e6f42d73e28\") " pod="openshift-infra/auto-csr-approver-29562732-s2frn" Mar 17 16:12:00 crc kubenswrapper[4767]: I0317 16:12:00.520806 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562732-s2frn" Mar 17 16:12:01 crc kubenswrapper[4767]: I0317 16:12:01.068083 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 16:12:01 crc kubenswrapper[4767]: I0317 16:12:01.074079 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562732-s2frn"] Mar 17 16:12:01 crc kubenswrapper[4767]: I0317 16:12:01.210199 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562732-s2frn" event={"ID":"1af290f7-a893-41d9-87ce-4e6f42d73e28","Type":"ContainerStarted","Data":"537bde54e55ee0dbd806c0e41280e5e956137e800569ddc80b12b9ff928fc95f"} Mar 17 16:12:02 crc kubenswrapper[4767]: I0317 16:12:02.054382 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4hdjg"] Mar 17 16:12:02 crc kubenswrapper[4767]: I0317 16:12:02.072495 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4hdjg"] Mar 17 16:12:03 crc kubenswrapper[4767]: I0317 16:12:03.245829 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562732-s2frn" event={"ID":"1af290f7-a893-41d9-87ce-4e6f42d73e28","Type":"ContainerStarted","Data":"e6cf33aa20c58d913e829a4d89bff396eb3e049dc1a45765a5d217387eebeb29"} Mar 17 16:12:03 crc kubenswrapper[4767]: I0317 16:12:03.275986 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562732-s2frn" podStartSLOduration=1.984519691 podStartE2EDuration="3.27595176s" podCreationTimestamp="2026-03-17 16:12:00 +0000 UTC" firstStartedPulling="2026-03-17 16:12:01.067872999 +0000 UTC m=+2112.481189046" lastFinishedPulling="2026-03-17 16:12:02.359305068 +0000 UTC m=+2113.772621115" observedRunningTime="2026-03-17 16:12:03.262072375 +0000 UTC m=+2114.675388442" watchObservedRunningTime="2026-03-17 16:12:03.27595176 +0000 UTC m=+2114.689267807" Mar 17 16:12:03 crc kubenswrapper[4767]: I0317 16:12:03.378688 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d3ec9f9-2461-4c46-afe3-f71f8b956cfa" path="/var/lib/kubelet/pods/9d3ec9f9-2461-4c46-afe3-f71f8b956cfa/volumes" Mar 17 16:12:04 crc kubenswrapper[4767]: I0317 16:12:04.166116 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:12:04 crc kubenswrapper[4767]: I0317 16:12:04.166201 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:12:04 crc kubenswrapper[4767]: I0317 16:12:04.282845 4767 generic.go:334] "Generic (PLEG): container finished" podID="1af290f7-a893-41d9-87ce-4e6f42d73e28" containerID="e6cf33aa20c58d913e829a4d89bff396eb3e049dc1a45765a5d217387eebeb29" exitCode=0 Mar 17 16:12:04 crc kubenswrapper[4767]: I0317 16:12:04.282927 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562732-s2frn" event={"ID":"1af290f7-a893-41d9-87ce-4e6f42d73e28","Type":"ContainerDied","Data":"e6cf33aa20c58d913e829a4d89bff396eb3e049dc1a45765a5d217387eebeb29"} Mar 17 16:12:05 crc kubenswrapper[4767]: I0317 16:12:05.756868 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562732-s2frn" Mar 17 16:12:05 crc kubenswrapper[4767]: I0317 16:12:05.952657 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxhqv\" (UniqueName: \"kubernetes.io/projected/1af290f7-a893-41d9-87ce-4e6f42d73e28-kube-api-access-bxhqv\") pod \"1af290f7-a893-41d9-87ce-4e6f42d73e28\" (UID: \"1af290f7-a893-41d9-87ce-4e6f42d73e28\") " Mar 17 16:12:05 crc kubenswrapper[4767]: I0317 16:12:05.962561 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af290f7-a893-41d9-87ce-4e6f42d73e28-kube-api-access-bxhqv" (OuterVolumeSpecName: "kube-api-access-bxhqv") pod "1af290f7-a893-41d9-87ce-4e6f42d73e28" (UID: "1af290f7-a893-41d9-87ce-4e6f42d73e28"). InnerVolumeSpecName "kube-api-access-bxhqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:12:06 crc kubenswrapper[4767]: I0317 16:12:06.057666 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxhqv\" (UniqueName: \"kubernetes.io/projected/1af290f7-a893-41d9-87ce-4e6f42d73e28-kube-api-access-bxhqv\") on node \"crc\" DevicePath \"\"" Mar 17 16:12:06 crc kubenswrapper[4767]: I0317 16:12:06.314584 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562732-s2frn" event={"ID":"1af290f7-a893-41d9-87ce-4e6f42d73e28","Type":"ContainerDied","Data":"537bde54e55ee0dbd806c0e41280e5e956137e800569ddc80b12b9ff928fc95f"} Mar 17 16:12:06 crc kubenswrapper[4767]: I0317 16:12:06.314922 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="537bde54e55ee0dbd806c0e41280e5e956137e800569ddc80b12b9ff928fc95f" Mar 17 16:12:06 crc kubenswrapper[4767]: I0317 16:12:06.314994 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562732-s2frn" Mar 17 16:12:06 crc kubenswrapper[4767]: I0317 16:12:06.349987 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562726-sq886"] Mar 17 16:12:06 crc kubenswrapper[4767]: I0317 16:12:06.384160 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562726-sq886"] Mar 17 16:12:07 crc kubenswrapper[4767]: I0317 16:12:07.368861 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54cf85c2-ceca-418b-a22e-ecadc2a3acfc" path="/var/lib/kubelet/pods/54cf85c2-ceca-418b-a22e-ecadc2a3acfc/volumes" Mar 17 16:12:10 crc kubenswrapper[4767]: I0317 16:12:10.042697 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c546-account-create-update-thtgc"] Mar 17 16:12:10 crc kubenswrapper[4767]: I0317 16:12:10.058507 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-mdmg7"] Mar 17 16:12:10 crc kubenswrapper[4767]: I0317 16:12:10.074326 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-0b49-account-create-update-gdff8"] Mar 17 16:12:10 crc kubenswrapper[4767]: I0317 16:12:10.086209 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-82a2-account-create-update-q7rxh"] Mar 17 16:12:10 crc kubenswrapper[4767]: I0317 16:12:10.096982 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-mdmg7"] Mar 17 16:12:10 crc kubenswrapper[4767]: I0317 16:12:10.108134 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c546-account-create-update-thtgc"] Mar 17 16:12:10 crc kubenswrapper[4767]: I0317 16:12:10.118870 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-82a2-account-create-update-q7rxh"] Mar 17 16:12:10 crc kubenswrapper[4767]: I0317 16:12:10.130005 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-0b49-account-create-update-gdff8"] Mar 17 16:12:11 crc kubenswrapper[4767]: I0317 16:12:11.679569 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0527b29f-1349-403b-86b4-8003c04e3925" path="/var/lib/kubelet/pods/0527b29f-1349-403b-86b4-8003c04e3925/volumes" Mar 17 16:12:11 crc kubenswrapper[4767]: I0317 16:12:11.681329 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64df007c-c140-48d9-9964-eff020f5d581" path="/var/lib/kubelet/pods/64df007c-c140-48d9-9964-eff020f5d581/volumes" Mar 17 16:12:11 crc kubenswrapper[4767]: I0317 16:12:11.690490 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6adfcc51-d4a8-451b-8685-4e822b1be5f9" path="/var/lib/kubelet/pods/6adfcc51-d4a8-451b-8685-4e822b1be5f9/volumes" Mar 17 16:12:11 crc kubenswrapper[4767]: I0317 16:12:11.694657 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94877582-ae44-41c9-8a66-1b7d323c71bb" path="/var/lib/kubelet/pods/94877582-ae44-41c9-8a66-1b7d323c71bb/volumes" Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.059941 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-2trqj"] Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.075314 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rsxsx"] Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.089070 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-883e-account-create-update-qh7zx"] Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.102387 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-58d46"] Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.114343 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rsxsx"] Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.126013 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-883e-account-create-update-qh7zx"] Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.138818 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-58d46"] Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.151962 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-2trqj"] Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.368335 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2107299-7bc5-4279-a443-7a266e21cec5" path="/var/lib/kubelet/pods/a2107299-7bc5-4279-a443-7a266e21cec5/volumes" Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.369640 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2" path="/var/lib/kubelet/pods/ac7c8a69-3ec6-440a-a4e8-5d52b2112ac2/volumes" Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.370786 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d532ab-b620-428e-a0ae-a4793930a703" path="/var/lib/kubelet/pods/b6d532ab-b620-428e-a0ae-a4793930a703/volumes" Mar 17 16:12:15 crc kubenswrapper[4767]: I0317 16:12:15.371493 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c678939f-b738-4c33-a849-175525de0122" path="/var/lib/kubelet/pods/c678939f-b738-4c33-a849-175525de0122/volumes" Mar 17 16:12:16 crc kubenswrapper[4767]: I0317 16:12:16.835581 4767 scope.go:117] "RemoveContainer" containerID="72a4f5442998b9192efa01ca5f0c3807d15f253aeefc3e81e59f202eae4f5c41" Mar 17 16:12:16 crc kubenswrapper[4767]: I0317 16:12:16.880130 4767 scope.go:117] "RemoveContainer" containerID="97d3145338f2803a261d84bfbb65c532f444c029b91eb8768f958fe6ee461023" Mar 17 16:12:16 crc kubenswrapper[4767]: I0317 16:12:16.946471 4767 scope.go:117] "RemoveContainer" containerID="0bd8254376f5c2166be6220136c28289e2241312119ecc0a6eae2580e67c364d" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.017615 4767 scope.go:117] "RemoveContainer" containerID="5a889d3abc4ff7270493d857f5dc795cb50a42f5bbe73708eae640967edb8669" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.093431 4767 scope.go:117] "RemoveContainer" containerID="07ab6f0deb52c5b01b03c33b2995e9757ba273c5e3b14e9cdb1d31e8cf2065bc" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.146247 4767 scope.go:117] "RemoveContainer" containerID="9b0e569d86178fd324ef0aaf639095da90f4eafa4e2367069824febfa9ecfe9e" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.200853 4767 scope.go:117] "RemoveContainer" containerID="a4fc72aebda01e02bd581b2042c114d5ae3b7b185f54c2c57855e9704e2932b8" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.233183 4767 scope.go:117] "RemoveContainer" containerID="e9a43787ac418a99df162e2c9d652dac1b1aecc999986ccde30a0737758453e1" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.288844 4767 scope.go:117] "RemoveContainer" containerID="4da7550950abbb01ac3501a84e75b92ab14da305bd43c38473c9eedeed31c059" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.319415 4767 scope.go:117] "RemoveContainer" containerID="a00da337f62baf20a19d907fed47f45f34e8040f172fcb2e0b78fabd9ceddb08" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.353084 4767 scope.go:117] "RemoveContainer" containerID="5c726a1f548330a53788b2877ffd12622841da12ca66e91210410e7b3d99fe87" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.389142 4767 scope.go:117] "RemoveContainer" containerID="91e2edd07554b3b5796895b4011d3d7141c61280aa1ac94d4e74b222b19dd295" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.423259 4767 scope.go:117] "RemoveContainer" containerID="d0e8afb8e4a04c05eeea84bd1c08a897ea98eb0dc1a8aa418836e6d3e9b015a2" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.452420 4767 scope.go:117] "RemoveContainer" containerID="25a5ddb7324c160d2823c76c7b6b889aded1ecf0cb07c04712f1d7c75412d7f4" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.481310 4767 scope.go:117] "RemoveContainer" containerID="50a8e92a1f7ae1f117fb00c47296a9dac97c1b41dda1b92092c3d4d4a35f537a" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.511024 4767 scope.go:117] "RemoveContainer" containerID="e2ddeff074fda785b4914c153886c6fe3788507e73308757b73327834debf484" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.538115 4767 scope.go:117] "RemoveContainer" containerID="c76e829586e52e8d8891839b90d6ee389c265ed0a0da0268cff0c9cf09ad579e" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.565968 4767 scope.go:117] "RemoveContainer" containerID="656cb70f5661b543ee3d2038ffd65436446a83cee9f1d0a01b17533d30f9b31e" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.594076 4767 scope.go:117] "RemoveContainer" containerID="d0d30ee80606f34c1edb9b3aae94618126577c6d6a4159354109d62d1860ac05" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.620004 4767 scope.go:117] "RemoveContainer" containerID="bf9587e52cd4d776af85f2888f638c71ef1d56c858ac4293357a4b8e18541c40" Mar 17 16:12:17 crc kubenswrapper[4767]: I0317 16:12:17.654401 4767 scope.go:117] "RemoveContainer" containerID="5a541f5e86b1f3f83a9bd446147a081f76ea3343987be804508edd1004ad4db8" Mar 17 16:12:20 crc kubenswrapper[4767]: I0317 16:12:20.075560 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-zbsv7"] Mar 17 16:12:20 crc kubenswrapper[4767]: I0317 16:12:20.088280 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-zbsv7"] Mar 17 16:12:21 crc kubenswrapper[4767]: I0317 16:12:21.371418 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6b00064-98f4-4f9d-82cb-a6aee03dcbeb" path="/var/lib/kubelet/pods/f6b00064-98f4-4f9d-82cb-a6aee03dcbeb/volumes" Mar 17 16:12:34 crc kubenswrapper[4767]: I0317 16:12:34.166416 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:12:34 crc kubenswrapper[4767]: I0317 16:12:34.167023 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:12:34 crc kubenswrapper[4767]: I0317 16:12:34.167083 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 16:12:34 crc kubenswrapper[4767]: I0317 16:12:34.168256 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"66ab96d70d85b8ad4af97b2cc9506406fc51b53c84a92e8d73017d6fc8cc0c7f"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 16:12:34 crc kubenswrapper[4767]: I0317 16:12:34.168334 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://66ab96d70d85b8ad4af97b2cc9506406fc51b53c84a92e8d73017d6fc8cc0c7f" gracePeriod=600 Mar 17 16:12:36 crc kubenswrapper[4767]: I0317 16:12:36.308793 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="66ab96d70d85b8ad4af97b2cc9506406fc51b53c84a92e8d73017d6fc8cc0c7f" exitCode=0 Mar 17 16:12:36 crc kubenswrapper[4767]: I0317 16:12:36.309425 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"66ab96d70d85b8ad4af97b2cc9506406fc51b53c84a92e8d73017d6fc8cc0c7f"} Mar 17 16:12:36 crc kubenswrapper[4767]: I0317 16:12:36.309469 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36"} Mar 17 16:12:36 crc kubenswrapper[4767]: I0317 16:12:36.309496 4767 scope.go:117] "RemoveContainer" containerID="0a380e88a442195fb2b5308595d8b016e9b6126f37577860a80fbc1dc81efe6c" Mar 17 16:13:01 crc kubenswrapper[4767]: I0317 16:13:01.051328 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-gm7c4"] Mar 17 16:13:01 crc kubenswrapper[4767]: I0317 16:13:01.064753 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-gm7c4"] Mar 17 16:13:01 crc kubenswrapper[4767]: I0317 16:13:01.371532 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="958adc8c-3c1c-4b35-8cc0-fa7a3b600842" path="/var/lib/kubelet/pods/958adc8c-3c1c-4b35-8cc0-fa7a3b600842/volumes" Mar 17 16:13:06 crc kubenswrapper[4767]: I0317 16:13:06.693461 4767 generic.go:334] "Generic (PLEG): container finished" podID="4e6c7649-a40f-44c0-8bcb-f4aa31251dcd" containerID="0066ca8ea83821688b5a754dce3b98eec35e082d143fdfda8902694395ff6c89" exitCode=0 Mar 17 16:13:06 crc kubenswrapper[4767]: I0317 16:13:06.693561 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" event={"ID":"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd","Type":"ContainerDied","Data":"0066ca8ea83821688b5a754dce3b98eec35e082d143fdfda8902694395ff6c89"} Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.343855 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.506612 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb5pp\" (UniqueName: \"kubernetes.io/projected/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-kube-api-access-bb5pp\") pod \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.506798 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-ssh-key-openstack-edpm-ipam\") pod \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.507006 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-bootstrap-combined-ca-bundle\") pod \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.507130 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-inventory\") pod \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\" (UID: \"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd\") " Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.514785 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "4e6c7649-a40f-44c0-8bcb-f4aa31251dcd" (UID: "4e6c7649-a40f-44c0-8bcb-f4aa31251dcd"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.527780 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-kube-api-access-bb5pp" (OuterVolumeSpecName: "kube-api-access-bb5pp") pod "4e6c7649-a40f-44c0-8bcb-f4aa31251dcd" (UID: "4e6c7649-a40f-44c0-8bcb-f4aa31251dcd"). InnerVolumeSpecName "kube-api-access-bb5pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.542996 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-inventory" (OuterVolumeSpecName: "inventory") pod "4e6c7649-a40f-44c0-8bcb-f4aa31251dcd" (UID: "4e6c7649-a40f-44c0-8bcb-f4aa31251dcd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.546714 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4e6c7649-a40f-44c0-8bcb-f4aa31251dcd" (UID: "4e6c7649-a40f-44c0-8bcb-f4aa31251dcd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.611122 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.611268 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb5pp\" (UniqueName: \"kubernetes.io/projected/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-kube-api-access-bb5pp\") on node \"crc\" DevicePath \"\"" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.611287 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.611298 4767 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6c7649-a40f-44c0-8bcb-f4aa31251dcd-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.723971 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" event={"ID":"4e6c7649-a40f-44c0-8bcb-f4aa31251dcd","Type":"ContainerDied","Data":"106e42f05f5c4d82538bcac32b72638009b70e4c13701852095714ea1d7352cf"} Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.724040 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="106e42f05f5c4d82538bcac32b72638009b70e4c13701852095714ea1d7352cf" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.724153 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-27ppt" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.827935 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp"] Mar 17 16:13:08 crc kubenswrapper[4767]: E0317 16:13:08.828913 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6c7649-a40f-44c0-8bcb-f4aa31251dcd" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.828949 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6c7649-a40f-44c0-8bcb-f4aa31251dcd" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 17 16:13:08 crc kubenswrapper[4767]: E0317 16:13:08.828976 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af290f7-a893-41d9-87ce-4e6f42d73e28" containerName="oc" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.828984 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af290f7-a893-41d9-87ce-4e6f42d73e28" containerName="oc" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.829428 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e6c7649-a40f-44c0-8bcb-f4aa31251dcd" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.829469 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af290f7-a893-41d9-87ce-4e6f42d73e28" containerName="oc" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.830830 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.833969 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.834013 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.834336 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.834466 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.854152 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp"] Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.920836 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.921456 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:13:08 crc kubenswrapper[4767]: I0317 16:13:08.921768 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp7hg\" (UniqueName: \"kubernetes.io/projected/9104d786-64f1-41a2-9f89-1863c3101edd-kube-api-access-qp7hg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:13:09 crc kubenswrapper[4767]: I0317 16:13:09.025256 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:13:09 crc kubenswrapper[4767]: I0317 16:13:09.025407 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp7hg\" (UniqueName: \"kubernetes.io/projected/9104d786-64f1-41a2-9f89-1863c3101edd-kube-api-access-qp7hg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:13:09 crc kubenswrapper[4767]: I0317 16:13:09.025489 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:13:09 crc kubenswrapper[4767]: I0317 16:13:09.030660 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:13:09 crc kubenswrapper[4767]: I0317 16:13:09.030661 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:13:09 crc kubenswrapper[4767]: I0317 16:13:09.046703 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp7hg\" (UniqueName: \"kubernetes.io/projected/9104d786-64f1-41a2-9f89-1863c3101edd-kube-api-access-qp7hg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:13:09 crc kubenswrapper[4767]: I0317 16:13:09.159272 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:13:09 crc kubenswrapper[4767]: I0317 16:13:09.715774 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp"] Mar 17 16:13:09 crc kubenswrapper[4767]: I0317 16:13:09.745707 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" event={"ID":"9104d786-64f1-41a2-9f89-1863c3101edd","Type":"ContainerStarted","Data":"288adf1682fa197151c2375f956d865ebe6e34f762776b6bb67eb900586165fb"} Mar 17 16:13:10 crc kubenswrapper[4767]: I0317 16:13:10.758923 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" event={"ID":"9104d786-64f1-41a2-9f89-1863c3101edd","Type":"ContainerStarted","Data":"b0f1a04e8d8770f7683f3c1b53ad891735b9c786f3413f43157c8d9ca9cd82bc"} Mar 17 16:13:10 crc kubenswrapper[4767]: I0317 16:13:10.792082 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" podStartSLOduration=2.35595109 podStartE2EDuration="2.792052986s" podCreationTimestamp="2026-03-17 16:13:08 +0000 UTC" firstStartedPulling="2026-03-17 16:13:09.72420642 +0000 UTC m=+2181.137522467" lastFinishedPulling="2026-03-17 16:13:10.160308316 +0000 UTC m=+2181.573624363" observedRunningTime="2026-03-17 16:13:10.788252228 +0000 UTC m=+2182.201568275" watchObservedRunningTime="2026-03-17 16:13:10.792052986 +0000 UTC m=+2182.205369023" Mar 17 16:13:18 crc kubenswrapper[4767]: I0317 16:13:18.137750 4767 scope.go:117] "RemoveContainer" containerID="30a882e8ff6de652846266fdbaa65c3ce1ede5bb1a5f6e4c2de39f25f6654fa9" Mar 17 16:13:18 crc kubenswrapper[4767]: I0317 16:13:18.169832 4767 scope.go:117] "RemoveContainer" containerID="52ab00d786ed6e40200bcdf22fa072d25c15f2ba3bec978aae02cda4cddbd0bb" Mar 17 16:13:18 crc kubenswrapper[4767]: I0317 16:13:18.212189 4767 scope.go:117] "RemoveContainer" containerID="0691bbc5bde42dfdec202493086bba02f881dbfe5ca0aee08ba6d976da9dfedb" Mar 17 16:13:18 crc kubenswrapper[4767]: I0317 16:13:18.264590 4767 scope.go:117] "RemoveContainer" containerID="a0f1af2de4965844a935ea59d92271222ff3b15093bb1f72eedb28d97f9e7e2b" Mar 17 16:13:18 crc kubenswrapper[4767]: I0317 16:13:18.306209 4767 scope.go:117] "RemoveContainer" containerID="207a85d187102e1a1442134ad96bbcd58337d2da80439bd774ba9ac839b6831e" Mar 17 16:13:23 crc kubenswrapper[4767]: I0317 16:13:23.080768 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-vhwq7"] Mar 17 16:13:23 crc kubenswrapper[4767]: I0317 16:13:23.093618 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-xj9pj"] Mar 17 16:13:23 crc kubenswrapper[4767]: I0317 16:13:23.108471 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-mzst5"] Mar 17 16:13:23 crc kubenswrapper[4767]: I0317 16:13:23.119527 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-vhwq7"] Mar 17 16:13:23 crc kubenswrapper[4767]: I0317 16:13:23.134495 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-mzst5"] Mar 17 16:13:23 crc kubenswrapper[4767]: I0317 16:13:23.147857 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-xj9pj"] Mar 17 16:13:23 crc kubenswrapper[4767]: I0317 16:13:23.370133 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="850bce7d-74fa-4ead-b6c2-fcf21c3083ec" path="/var/lib/kubelet/pods/850bce7d-74fa-4ead-b6c2-fcf21c3083ec/volumes" Mar 17 16:13:23 crc kubenswrapper[4767]: I0317 16:13:23.372170 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5b58621-27a1-479d-a90c-aa96400e1116" path="/var/lib/kubelet/pods/c5b58621-27a1-479d-a90c-aa96400e1116/volumes" Mar 17 16:13:23 crc kubenswrapper[4767]: I0317 16:13:23.373643 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2af7040-49bc-4689-8af6-a7e3a37db7f2" path="/var/lib/kubelet/pods/f2af7040-49bc-4689-8af6-a7e3a37db7f2/volumes" Mar 17 16:13:48 crc kubenswrapper[4767]: I0317 16:13:48.052627 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-pdxns"] Mar 17 16:13:48 crc kubenswrapper[4767]: I0317 16:13:48.074541 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-pdxns"] Mar 17 16:13:49 crc kubenswrapper[4767]: I0317 16:13:49.370447 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="712803ce-6e47-40a6-bf5c-84e8aea748e4" path="/var/lib/kubelet/pods/712803ce-6e47-40a6-bf5c-84e8aea748e4/volumes" Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.738428 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bstjm"] Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.742053 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.753046 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bstjm"] Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.787548 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-catalog-content\") pod \"community-operators-bstjm\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.787711 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-utilities\") pod \"community-operators-bstjm\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.787824 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnlvb\" (UniqueName: \"kubernetes.io/projected/2d96ed67-5e7d-47e1-870e-4156599b0e07-kube-api-access-mnlvb\") pod \"community-operators-bstjm\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.891662 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-utilities\") pod \"community-operators-bstjm\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.891824 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnlvb\" (UniqueName: \"kubernetes.io/projected/2d96ed67-5e7d-47e1-870e-4156599b0e07-kube-api-access-mnlvb\") pod \"community-operators-bstjm\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.892141 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-catalog-content\") pod \"community-operators-bstjm\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.892800 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-catalog-content\") pod \"community-operators-bstjm\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.893058 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-utilities\") pod \"community-operators-bstjm\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:13:58 crc kubenswrapper[4767]: I0317 16:13:58.918400 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnlvb\" (UniqueName: \"kubernetes.io/projected/2d96ed67-5e7d-47e1-870e-4156599b0e07-kube-api-access-mnlvb\") pod \"community-operators-bstjm\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:13:59 crc kubenswrapper[4767]: I0317 16:13:59.075668 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.155445 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562734-chbkz"] Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.158864 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562734-chbkz" Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.170887 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562734-chbkz"] Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.171924 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.171950 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.172105 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.203085 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bstjm"] Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.251914 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4glcq\" (UniqueName: \"kubernetes.io/projected/18fa314c-78e8-4e3e-a392-855c08a28f62-kube-api-access-4glcq\") pod \"auto-csr-approver-29562734-chbkz\" (UID: \"18fa314c-78e8-4e3e-a392-855c08a28f62\") " pod="openshift-infra/auto-csr-approver-29562734-chbkz" Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.355816 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4glcq\" (UniqueName: \"kubernetes.io/projected/18fa314c-78e8-4e3e-a392-855c08a28f62-kube-api-access-4glcq\") pod \"auto-csr-approver-29562734-chbkz\" (UID: \"18fa314c-78e8-4e3e-a392-855c08a28f62\") " pod="openshift-infra/auto-csr-approver-29562734-chbkz" Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.382899 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4glcq\" (UniqueName: \"kubernetes.io/projected/18fa314c-78e8-4e3e-a392-855c08a28f62-kube-api-access-4glcq\") pod \"auto-csr-approver-29562734-chbkz\" (UID: \"18fa314c-78e8-4e3e-a392-855c08a28f62\") " pod="openshift-infra/auto-csr-approver-29562734-chbkz" Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.411643 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstjm" event={"ID":"2d96ed67-5e7d-47e1-870e-4156599b0e07","Type":"ContainerStarted","Data":"e4dc9018e73ae3f8f7bb829e87ee9daace2da34c117a8d14771b8296f8c83948"} Mar 17 16:14:00 crc kubenswrapper[4767]: I0317 16:14:00.489333 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562734-chbkz" Mar 17 16:14:01 crc kubenswrapper[4767]: W0317 16:14:01.002070 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18fa314c_78e8_4e3e_a392_855c08a28f62.slice/crio-5e2295ef6da288d283ae24e2309f806fcfb290719833a50a27e44ec88341775b WatchSource:0}: Error finding container 5e2295ef6da288d283ae24e2309f806fcfb290719833a50a27e44ec88341775b: Status 404 returned error can't find the container with id 5e2295ef6da288d283ae24e2309f806fcfb290719833a50a27e44ec88341775b Mar 17 16:14:01 crc kubenswrapper[4767]: I0317 16:14:01.002082 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562734-chbkz"] Mar 17 16:14:01 crc kubenswrapper[4767]: I0317 16:14:01.425427 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562734-chbkz" event={"ID":"18fa314c-78e8-4e3e-a392-855c08a28f62","Type":"ContainerStarted","Data":"5e2295ef6da288d283ae24e2309f806fcfb290719833a50a27e44ec88341775b"} Mar 17 16:14:01 crc kubenswrapper[4767]: I0317 16:14:01.429975 4767 generic.go:334] "Generic (PLEG): container finished" podID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerID="72b2038e589267138cdf368688709bb8845fc1e61b60bf732661eebfb59c7e6c" exitCode=0 Mar 17 16:14:01 crc kubenswrapper[4767]: I0317 16:14:01.430033 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstjm" event={"ID":"2d96ed67-5e7d-47e1-870e-4156599b0e07","Type":"ContainerDied","Data":"72b2038e589267138cdf368688709bb8845fc1e61b60bf732661eebfb59c7e6c"} Mar 17 16:14:03 crc kubenswrapper[4767]: I0317 16:14:03.471456 4767 generic.go:334] "Generic (PLEG): container finished" podID="18fa314c-78e8-4e3e-a392-855c08a28f62" containerID="62f8005b8159d57801c533da8eebe0dbed1db72194cbe800b73591ac42cb22b9" exitCode=0 Mar 17 16:14:03 crc kubenswrapper[4767]: I0317 16:14:03.472092 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562734-chbkz" event={"ID":"18fa314c-78e8-4e3e-a392-855c08a28f62","Type":"ContainerDied","Data":"62f8005b8159d57801c533da8eebe0dbed1db72194cbe800b73591ac42cb22b9"} Mar 17 16:14:03 crc kubenswrapper[4767]: I0317 16:14:03.476052 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstjm" event={"ID":"2d96ed67-5e7d-47e1-870e-4156599b0e07","Type":"ContainerStarted","Data":"07e7319ece75e24e695118d66afe60e6535749a0cf66b370b4ca3df5648861c2"} Mar 17 16:14:04 crc kubenswrapper[4767]: I0317 16:14:04.514644 4767 generic.go:334] "Generic (PLEG): container finished" podID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerID="07e7319ece75e24e695118d66afe60e6535749a0cf66b370b4ca3df5648861c2" exitCode=0 Mar 17 16:14:04 crc kubenswrapper[4767]: I0317 16:14:04.514724 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstjm" event={"ID":"2d96ed67-5e7d-47e1-870e-4156599b0e07","Type":"ContainerDied","Data":"07e7319ece75e24e695118d66afe60e6535749a0cf66b370b4ca3df5648861c2"} Mar 17 16:14:05 crc kubenswrapper[4767]: I0317 16:14:05.046793 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562734-chbkz" Mar 17 16:14:05 crc kubenswrapper[4767]: I0317 16:14:05.131764 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4glcq\" (UniqueName: \"kubernetes.io/projected/18fa314c-78e8-4e3e-a392-855c08a28f62-kube-api-access-4glcq\") pod \"18fa314c-78e8-4e3e-a392-855c08a28f62\" (UID: \"18fa314c-78e8-4e3e-a392-855c08a28f62\") " Mar 17 16:14:05 crc kubenswrapper[4767]: I0317 16:14:05.137602 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18fa314c-78e8-4e3e-a392-855c08a28f62-kube-api-access-4glcq" (OuterVolumeSpecName: "kube-api-access-4glcq") pod "18fa314c-78e8-4e3e-a392-855c08a28f62" (UID: "18fa314c-78e8-4e3e-a392-855c08a28f62"). InnerVolumeSpecName "kube-api-access-4glcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:14:05 crc kubenswrapper[4767]: I0317 16:14:05.236945 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4glcq\" (UniqueName: \"kubernetes.io/projected/18fa314c-78e8-4e3e-a392-855c08a28f62-kube-api-access-4glcq\") on node \"crc\" DevicePath \"\"" Mar 17 16:14:05 crc kubenswrapper[4767]: I0317 16:14:05.532125 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562734-chbkz" Mar 17 16:14:05 crc kubenswrapper[4767]: I0317 16:14:05.532126 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562734-chbkz" event={"ID":"18fa314c-78e8-4e3e-a392-855c08a28f62","Type":"ContainerDied","Data":"5e2295ef6da288d283ae24e2309f806fcfb290719833a50a27e44ec88341775b"} Mar 17 16:14:05 crc kubenswrapper[4767]: I0317 16:14:05.532214 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e2295ef6da288d283ae24e2309f806fcfb290719833a50a27e44ec88341775b" Mar 17 16:14:05 crc kubenswrapper[4767]: I0317 16:14:05.535324 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstjm" event={"ID":"2d96ed67-5e7d-47e1-870e-4156599b0e07","Type":"ContainerStarted","Data":"519e3b1bdeeaa4c673f5bd05329388ab877fd7b5f70438454343a3f7a6cb443a"} Mar 17 16:14:05 crc kubenswrapper[4767]: I0317 16:14:05.559129 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bstjm" podStartSLOduration=4.001206632 podStartE2EDuration="7.559105076s" podCreationTimestamp="2026-03-17 16:13:58 +0000 UTC" firstStartedPulling="2026-03-17 16:14:01.43259154 +0000 UTC m=+2232.845907587" lastFinishedPulling="2026-03-17 16:14:04.990489984 +0000 UTC m=+2236.403806031" observedRunningTime="2026-03-17 16:14:05.557297674 +0000 UTC m=+2236.970613741" watchObservedRunningTime="2026-03-17 16:14:05.559105076 +0000 UTC m=+2236.972421143" Mar 17 16:14:06 crc kubenswrapper[4767]: I0317 16:14:06.128295 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562728-g6ss4"] Mar 17 16:14:06 crc kubenswrapper[4767]: I0317 16:14:06.146043 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562728-g6ss4"] Mar 17 16:14:07 crc kubenswrapper[4767]: I0317 16:14:07.371017 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68a9bbf0-4d9d-4760-9aa7-3afc7aa20123" path="/var/lib/kubelet/pods/68a9bbf0-4d9d-4760-9aa7-3afc7aa20123/volumes" Mar 17 16:14:09 crc kubenswrapper[4767]: I0317 16:14:09.076053 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:14:09 crc kubenswrapper[4767]: I0317 16:14:09.077244 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:14:10 crc kubenswrapper[4767]: I0317 16:14:10.128040 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-bstjm" podUID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerName="registry-server" probeResult="failure" output=< Mar 17 16:14:10 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:14:10 crc kubenswrapper[4767]: > Mar 17 16:14:18 crc kubenswrapper[4767]: I0317 16:14:18.487067 4767 scope.go:117] "RemoveContainer" containerID="d83d0df8dc6726bc21c1a068465bd820f812594d7608c6a517469016f7b477dd" Mar 17 16:14:18 crc kubenswrapper[4767]: I0317 16:14:18.522421 4767 scope.go:117] "RemoveContainer" containerID="2fb9a9b45b63e921103b8b488be5012d01d9d2120233755ddea3cbee0a4f05b4" Mar 17 16:14:18 crc kubenswrapper[4767]: I0317 16:14:18.584619 4767 scope.go:117] "RemoveContainer" containerID="71cee4369dbc400c2680a693359d1227aa47dda82465b728603476ef258ef922" Mar 17 16:14:18 crc kubenswrapper[4767]: I0317 16:14:18.674118 4767 scope.go:117] "RemoveContainer" containerID="4b19420f7562b7140f30f8aa7b439c3e38ff297bd754e956e70d52d1805782e0" Mar 17 16:14:18 crc kubenswrapper[4767]: I0317 16:14:18.705030 4767 scope.go:117] "RemoveContainer" containerID="9e8adf44c914c054494ac45c8245d85c8be3816851cea224cae9508ce2b26587" Mar 17 16:14:19 crc kubenswrapper[4767]: I0317 16:14:19.147198 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:14:19 crc kubenswrapper[4767]: I0317 16:14:19.238836 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.301472 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tvff9"] Mar 17 16:14:21 crc kubenswrapper[4767]: E0317 16:14:21.303409 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18fa314c-78e8-4e3e-a392-855c08a28f62" containerName="oc" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.303439 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="18fa314c-78e8-4e3e-a392-855c08a28f62" containerName="oc" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.303858 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="18fa314c-78e8-4e3e-a392-855c08a28f62" containerName="oc" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.307220 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.318426 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvff9"] Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.337577 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-catalog-content\") pod \"redhat-marketplace-tvff9\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.337637 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtgtv\" (UniqueName: \"kubernetes.io/projected/25fab369-f3b3-40d4-8147-d71bf189aebd-kube-api-access-xtgtv\") pod \"redhat-marketplace-tvff9\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.337938 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-utilities\") pod \"redhat-marketplace-tvff9\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.447857 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-utilities\") pod \"redhat-marketplace-tvff9\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.448196 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-catalog-content\") pod \"redhat-marketplace-tvff9\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.448242 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtgtv\" (UniqueName: \"kubernetes.io/projected/25fab369-f3b3-40d4-8147-d71bf189aebd-kube-api-access-xtgtv\") pod \"redhat-marketplace-tvff9\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.449036 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-catalog-content\") pod \"redhat-marketplace-tvff9\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.449093 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-utilities\") pod \"redhat-marketplace-tvff9\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.478100 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtgtv\" (UniqueName: \"kubernetes.io/projected/25fab369-f3b3-40d4-8147-d71bf189aebd-kube-api-access-xtgtv\") pod \"redhat-marketplace-tvff9\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:21 crc kubenswrapper[4767]: I0317 16:14:21.642993 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:22 crc kubenswrapper[4767]: I0317 16:14:22.190750 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvff9"] Mar 17 16:14:22 crc kubenswrapper[4767]: I0317 16:14:22.950237 4767 generic.go:334] "Generic (PLEG): container finished" podID="25fab369-f3b3-40d4-8147-d71bf189aebd" containerID="35f37758832474f3bda7c80f1d0c31f930bd25f7f993246f72391a9f0bc5bd76" exitCode=0 Mar 17 16:14:22 crc kubenswrapper[4767]: I0317 16:14:22.950370 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvff9" event={"ID":"25fab369-f3b3-40d4-8147-d71bf189aebd","Type":"ContainerDied","Data":"35f37758832474f3bda7c80f1d0c31f930bd25f7f993246f72391a9f0bc5bd76"} Mar 17 16:14:22 crc kubenswrapper[4767]: I0317 16:14:22.950611 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvff9" event={"ID":"25fab369-f3b3-40d4-8147-d71bf189aebd","Type":"ContainerStarted","Data":"7eea19380f666b5599345b82e1e1728390dd5bea77fb675d384c5af2879a3d4c"} Mar 17 16:14:23 crc kubenswrapper[4767]: I0317 16:14:23.693692 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bstjm"] Mar 17 16:14:23 crc kubenswrapper[4767]: I0317 16:14:23.693999 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bstjm" podUID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerName="registry-server" containerID="cri-o://519e3b1bdeeaa4c673f5bd05329388ab877fd7b5f70438454343a3f7a6cb443a" gracePeriod=2 Mar 17 16:14:23 crc kubenswrapper[4767]: I0317 16:14:23.973648 4767 generic.go:334] "Generic (PLEG): container finished" podID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerID="519e3b1bdeeaa4c673f5bd05329388ab877fd7b5f70438454343a3f7a6cb443a" exitCode=0 Mar 17 16:14:23 crc kubenswrapper[4767]: I0317 16:14:23.973892 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstjm" event={"ID":"2d96ed67-5e7d-47e1-870e-4156599b0e07","Type":"ContainerDied","Data":"519e3b1bdeeaa4c673f5bd05329388ab877fd7b5f70438454343a3f7a6cb443a"} Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.475519 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.649138 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnlvb\" (UniqueName: \"kubernetes.io/projected/2d96ed67-5e7d-47e1-870e-4156599b0e07-kube-api-access-mnlvb\") pod \"2d96ed67-5e7d-47e1-870e-4156599b0e07\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.649713 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-catalog-content\") pod \"2d96ed67-5e7d-47e1-870e-4156599b0e07\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.649785 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-utilities\") pod \"2d96ed67-5e7d-47e1-870e-4156599b0e07\" (UID: \"2d96ed67-5e7d-47e1-870e-4156599b0e07\") " Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.650556 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-utilities" (OuterVolumeSpecName: "utilities") pod "2d96ed67-5e7d-47e1-870e-4156599b0e07" (UID: "2d96ed67-5e7d-47e1-870e-4156599b0e07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.652125 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.656642 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d96ed67-5e7d-47e1-870e-4156599b0e07-kube-api-access-mnlvb" (OuterVolumeSpecName: "kube-api-access-mnlvb") pod "2d96ed67-5e7d-47e1-870e-4156599b0e07" (UID: "2d96ed67-5e7d-47e1-870e-4156599b0e07"). InnerVolumeSpecName "kube-api-access-mnlvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.722707 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d96ed67-5e7d-47e1-870e-4156599b0e07" (UID: "2d96ed67-5e7d-47e1-870e-4156599b0e07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.755260 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnlvb\" (UniqueName: \"kubernetes.io/projected/2d96ed67-5e7d-47e1-870e-4156599b0e07-kube-api-access-mnlvb\") on node \"crc\" DevicePath \"\"" Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.755307 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d96ed67-5e7d-47e1-870e-4156599b0e07-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.988637 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bstjm" Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.988666 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstjm" event={"ID":"2d96ed67-5e7d-47e1-870e-4156599b0e07","Type":"ContainerDied","Data":"e4dc9018e73ae3f8f7bb829e87ee9daace2da34c117a8d14771b8296f8c83948"} Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.990208 4767 scope.go:117] "RemoveContainer" containerID="519e3b1bdeeaa4c673f5bd05329388ab877fd7b5f70438454343a3f7a6cb443a" Mar 17 16:14:24 crc kubenswrapper[4767]: I0317 16:14:24.991699 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvff9" event={"ID":"25fab369-f3b3-40d4-8147-d71bf189aebd","Type":"ContainerStarted","Data":"122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1"} Mar 17 16:14:25 crc kubenswrapper[4767]: I0317 16:14:25.028649 4767 scope.go:117] "RemoveContainer" containerID="07e7319ece75e24e695118d66afe60e6535749a0cf66b370b4ca3df5648861c2" Mar 17 16:14:25 crc kubenswrapper[4767]: I0317 16:14:25.060817 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bstjm"] Mar 17 16:14:25 crc kubenswrapper[4767]: I0317 16:14:25.065919 4767 scope.go:117] "RemoveContainer" containerID="72b2038e589267138cdf368688709bb8845fc1e61b60bf732661eebfb59c7e6c" Mar 17 16:14:25 crc kubenswrapper[4767]: I0317 16:14:25.074586 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bstjm"] Mar 17 16:14:25 crc kubenswrapper[4767]: I0317 16:14:25.374124 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d96ed67-5e7d-47e1-870e-4156599b0e07" path="/var/lib/kubelet/pods/2d96ed67-5e7d-47e1-870e-4156599b0e07/volumes" Mar 17 16:14:26 crc kubenswrapper[4767]: I0317 16:14:26.007808 4767 generic.go:334] "Generic (PLEG): container finished" podID="25fab369-f3b3-40d4-8147-d71bf189aebd" containerID="122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1" exitCode=0 Mar 17 16:14:26 crc kubenswrapper[4767]: I0317 16:14:26.007879 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvff9" event={"ID":"25fab369-f3b3-40d4-8147-d71bf189aebd","Type":"ContainerDied","Data":"122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1"} Mar 17 16:14:27 crc kubenswrapper[4767]: I0317 16:14:27.026460 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvff9" event={"ID":"25fab369-f3b3-40d4-8147-d71bf189aebd","Type":"ContainerStarted","Data":"ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00"} Mar 17 16:14:27 crc kubenswrapper[4767]: I0317 16:14:27.055609 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tvff9" podStartSLOduration=2.385412146 podStartE2EDuration="6.055586239s" podCreationTimestamp="2026-03-17 16:14:21 +0000 UTC" firstStartedPulling="2026-03-17 16:14:22.954039725 +0000 UTC m=+2254.367355772" lastFinishedPulling="2026-03-17 16:14:26.624213808 +0000 UTC m=+2258.037529865" observedRunningTime="2026-03-17 16:14:27.050668139 +0000 UTC m=+2258.463984186" watchObservedRunningTime="2026-03-17 16:14:27.055586239 +0000 UTC m=+2258.468902286" Mar 17 16:14:31 crc kubenswrapper[4767]: I0317 16:14:31.643616 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:31 crc kubenswrapper[4767]: I0317 16:14:31.644367 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:31 crc kubenswrapper[4767]: I0317 16:14:31.700679 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:32 crc kubenswrapper[4767]: I0317 16:14:32.144298 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:32 crc kubenswrapper[4767]: I0317 16:14:32.210474 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvff9"] Mar 17 16:14:34 crc kubenswrapper[4767]: I0317 16:14:34.114915 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tvff9" podUID="25fab369-f3b3-40d4-8147-d71bf189aebd" containerName="registry-server" containerID="cri-o://ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00" gracePeriod=2 Mar 17 16:14:34 crc kubenswrapper[4767]: I0317 16:14:34.748532 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:34 crc kubenswrapper[4767]: I0317 16:14:34.881116 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtgtv\" (UniqueName: \"kubernetes.io/projected/25fab369-f3b3-40d4-8147-d71bf189aebd-kube-api-access-xtgtv\") pod \"25fab369-f3b3-40d4-8147-d71bf189aebd\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " Mar 17 16:14:34 crc kubenswrapper[4767]: I0317 16:14:34.881748 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-catalog-content\") pod \"25fab369-f3b3-40d4-8147-d71bf189aebd\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " Mar 17 16:14:34 crc kubenswrapper[4767]: I0317 16:14:34.881950 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-utilities\") pod \"25fab369-f3b3-40d4-8147-d71bf189aebd\" (UID: \"25fab369-f3b3-40d4-8147-d71bf189aebd\") " Mar 17 16:14:34 crc kubenswrapper[4767]: I0317 16:14:34.882761 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-utilities" (OuterVolumeSpecName: "utilities") pod "25fab369-f3b3-40d4-8147-d71bf189aebd" (UID: "25fab369-f3b3-40d4-8147-d71bf189aebd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:14:34 crc kubenswrapper[4767]: I0317 16:14:34.897656 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25fab369-f3b3-40d4-8147-d71bf189aebd-kube-api-access-xtgtv" (OuterVolumeSpecName: "kube-api-access-xtgtv") pod "25fab369-f3b3-40d4-8147-d71bf189aebd" (UID: "25fab369-f3b3-40d4-8147-d71bf189aebd"). InnerVolumeSpecName "kube-api-access-xtgtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:14:34 crc kubenswrapper[4767]: I0317 16:14:34.916315 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25fab369-f3b3-40d4-8147-d71bf189aebd" (UID: "25fab369-f3b3-40d4-8147-d71bf189aebd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:14:34 crc kubenswrapper[4767]: I0317 16:14:34.985841 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:14:34 crc kubenswrapper[4767]: I0317 16:14:34.985891 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25fab369-f3b3-40d4-8147-d71bf189aebd-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:14:34 crc kubenswrapper[4767]: I0317 16:14:34.985907 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtgtv\" (UniqueName: \"kubernetes.io/projected/25fab369-f3b3-40d4-8147-d71bf189aebd-kube-api-access-xtgtv\") on node \"crc\" DevicePath \"\"" Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.129334 4767 generic.go:334] "Generic (PLEG): container finished" podID="25fab369-f3b3-40d4-8147-d71bf189aebd" containerID="ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00" exitCode=0 Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.129394 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvff9" event={"ID":"25fab369-f3b3-40d4-8147-d71bf189aebd","Type":"ContainerDied","Data":"ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00"} Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.129408 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvff9" Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.129449 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvff9" event={"ID":"25fab369-f3b3-40d4-8147-d71bf189aebd","Type":"ContainerDied","Data":"7eea19380f666b5599345b82e1e1728390dd5bea77fb675d384c5af2879a3d4c"} Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.129478 4767 scope.go:117] "RemoveContainer" containerID="ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00" Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.174076 4767 scope.go:117] "RemoveContainer" containerID="122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1" Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.180143 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvff9"] Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.191768 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvff9"] Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.203001 4767 scope.go:117] "RemoveContainer" containerID="35f37758832474f3bda7c80f1d0c31f930bd25f7f993246f72391a9f0bc5bd76" Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.267017 4767 scope.go:117] "RemoveContainer" containerID="ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00" Mar 17 16:14:35 crc kubenswrapper[4767]: E0317 16:14:35.267868 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00\": container with ID starting with ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00 not found: ID does not exist" containerID="ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00" Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.267930 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00"} err="failed to get container status \"ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00\": rpc error: code = NotFound desc = could not find container \"ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00\": container with ID starting with ba87084f3b5c12543ca8875721bf14b5f01d734f7fad8fde497e5cb236fc5a00 not found: ID does not exist" Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.267966 4767 scope.go:117] "RemoveContainer" containerID="122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1" Mar 17 16:14:35 crc kubenswrapper[4767]: E0317 16:14:35.268905 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1\": container with ID starting with 122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1 not found: ID does not exist" containerID="122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1" Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.268993 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1"} err="failed to get container status \"122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1\": rpc error: code = NotFound desc = could not find container \"122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1\": container with ID starting with 122b4b6ceca6f2d39bf5aa36952b63462034ef701842bb2b92f9b3a64c324dc1 not found: ID does not exist" Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.269065 4767 scope.go:117] "RemoveContainer" containerID="35f37758832474f3bda7c80f1d0c31f930bd25f7f993246f72391a9f0bc5bd76" Mar 17 16:14:35 crc kubenswrapper[4767]: E0317 16:14:35.271847 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35f37758832474f3bda7c80f1d0c31f930bd25f7f993246f72391a9f0bc5bd76\": container with ID starting with 35f37758832474f3bda7c80f1d0c31f930bd25f7f993246f72391a9f0bc5bd76 not found: ID does not exist" containerID="35f37758832474f3bda7c80f1d0c31f930bd25f7f993246f72391a9f0bc5bd76" Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.271913 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35f37758832474f3bda7c80f1d0c31f930bd25f7f993246f72391a9f0bc5bd76"} err="failed to get container status \"35f37758832474f3bda7c80f1d0c31f930bd25f7f993246f72391a9f0bc5bd76\": rpc error: code = NotFound desc = could not find container \"35f37758832474f3bda7c80f1d0c31f930bd25f7f993246f72391a9f0bc5bd76\": container with ID starting with 35f37758832474f3bda7c80f1d0c31f930bd25f7f993246f72391a9f0bc5bd76 not found: ID does not exist" Mar 17 16:14:35 crc kubenswrapper[4767]: I0317 16:14:35.372766 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25fab369-f3b3-40d4-8147-d71bf189aebd" path="/var/lib/kubelet/pods/25fab369-f3b3-40d4-8147-d71bf189aebd/volumes" Mar 17 16:14:54 crc kubenswrapper[4767]: I0317 16:14:54.048363 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-2vgbq"] Mar 17 16:14:54 crc kubenswrapper[4767]: I0317 16:14:54.063388 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-gjbh4"] Mar 17 16:14:54 crc kubenswrapper[4767]: I0317 16:14:54.078573 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-54a5-account-create-update-qzw6x"] Mar 17 16:14:54 crc kubenswrapper[4767]: I0317 16:14:54.090527 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-2vgbq"] Mar 17 16:14:54 crc kubenswrapper[4767]: I0317 16:14:54.103451 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-54a5-account-create-update-qzw6x"] Mar 17 16:14:54 crc kubenswrapper[4767]: I0317 16:14:54.114477 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-gjbh4"] Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.038758 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ed4a-account-create-update-49dhk"] Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.051084 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-f1bf-account-create-update-w5cxr"] Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.063723 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-5zdrh"] Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.076576 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ed4a-account-create-update-49dhk"] Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.089564 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-5zdrh"] Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.102468 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-f1bf-account-create-update-w5cxr"] Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.370441 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24e92570-fb4d-4fa2-ae4f-cc11da8f764a" path="/var/lib/kubelet/pods/24e92570-fb4d-4fa2-ae4f-cc11da8f764a/volumes" Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.371407 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="285fe2d6-1df4-4412-a107-a12bcb38e91f" path="/var/lib/kubelet/pods/285fe2d6-1df4-4412-a107-a12bcb38e91f/volumes" Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.372345 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63e8a6ca-f669-4eef-8509-f30f7ee31447" path="/var/lib/kubelet/pods/63e8a6ca-f669-4eef-8509-f30f7ee31447/volumes" Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.372997 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1137107-c196-4307-b2c4-90a6e9a0aab4" path="/var/lib/kubelet/pods/c1137107-c196-4307-b2c4-90a6e9a0aab4/volumes" Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.374249 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9a6a8fc-9770-4e16-8f14-371b008d3742" path="/var/lib/kubelet/pods/d9a6a8fc-9770-4e16-8f14-371b008d3742/volumes" Mar 17 16:14:55 crc kubenswrapper[4767]: I0317 16:14:55.374946 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efb414e8-1117-4306-85b8-311102e2d8fb" path="/var/lib/kubelet/pods/efb414e8-1117-4306-85b8-311102e2d8fb/volumes" Mar 17 16:14:56 crc kubenswrapper[4767]: I0317 16:14:56.945060 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.164396 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl"] Mar 17 16:15:00 crc kubenswrapper[4767]: E0317 16:15:00.165745 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerName="extract-content" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.165765 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerName="extract-content" Mar 17 16:15:00 crc kubenswrapper[4767]: E0317 16:15:00.165810 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25fab369-f3b3-40d4-8147-d71bf189aebd" containerName="extract-utilities" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.165824 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="25fab369-f3b3-40d4-8147-d71bf189aebd" containerName="extract-utilities" Mar 17 16:15:00 crc kubenswrapper[4767]: E0317 16:15:00.165858 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerName="registry-server" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.165867 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerName="registry-server" Mar 17 16:15:00 crc kubenswrapper[4767]: E0317 16:15:00.165882 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25fab369-f3b3-40d4-8147-d71bf189aebd" containerName="registry-server" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.165889 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="25fab369-f3b3-40d4-8147-d71bf189aebd" containerName="registry-server" Mar 17 16:15:00 crc kubenswrapper[4767]: E0317 16:15:00.165903 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerName="extract-utilities" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.165909 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerName="extract-utilities" Mar 17 16:15:00 crc kubenswrapper[4767]: E0317 16:15:00.165929 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25fab369-f3b3-40d4-8147-d71bf189aebd" containerName="extract-content" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.165937 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="25fab369-f3b3-40d4-8147-d71bf189aebd" containerName="extract-content" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.166291 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="25fab369-f3b3-40d4-8147-d71bf189aebd" containerName="registry-server" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.166320 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d96ed67-5e7d-47e1-870e-4156599b0e07" containerName="registry-server" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.167667 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.170723 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.171740 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.177243 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl"] Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.214501 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qhcv\" (UniqueName: \"kubernetes.io/projected/7ab4af60-5262-4a85-b749-961ca0b7b162-kube-api-access-4qhcv\") pod \"collect-profiles-29562735-z7fsl\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.214794 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ab4af60-5262-4a85-b749-961ca0b7b162-secret-volume\") pod \"collect-profiles-29562735-z7fsl\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.214854 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ab4af60-5262-4a85-b749-961ca0b7b162-config-volume\") pod \"collect-profiles-29562735-z7fsl\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.316493 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qhcv\" (UniqueName: \"kubernetes.io/projected/7ab4af60-5262-4a85-b749-961ca0b7b162-kube-api-access-4qhcv\") pod \"collect-profiles-29562735-z7fsl\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.316708 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ab4af60-5262-4a85-b749-961ca0b7b162-secret-volume\") pod \"collect-profiles-29562735-z7fsl\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.316756 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ab4af60-5262-4a85-b749-961ca0b7b162-config-volume\") pod \"collect-profiles-29562735-z7fsl\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.317820 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ab4af60-5262-4a85-b749-961ca0b7b162-config-volume\") pod \"collect-profiles-29562735-z7fsl\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.324385 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ab4af60-5262-4a85-b749-961ca0b7b162-secret-volume\") pod \"collect-profiles-29562735-z7fsl\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.335452 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qhcv\" (UniqueName: \"kubernetes.io/projected/7ab4af60-5262-4a85-b749-961ca0b7b162-kube-api-access-4qhcv\") pod \"collect-profiles-29562735-z7fsl\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.492726 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:00 crc kubenswrapper[4767]: I0317 16:15:00.991802 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl"] Mar 17 16:15:01 crc kubenswrapper[4767]: I0317 16:15:01.474036 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" event={"ID":"7ab4af60-5262-4a85-b749-961ca0b7b162","Type":"ContainerStarted","Data":"881f1867e56ab09c9385353926ad2b79d8af28c1cae6ef12f5934ce58d1608b2"} Mar 17 16:15:01 crc kubenswrapper[4767]: I0317 16:15:01.474438 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" event={"ID":"7ab4af60-5262-4a85-b749-961ca0b7b162","Type":"ContainerStarted","Data":"4572ae94a921ca68520ee0446c557dc08989b5109e94b1562b33c05ba8d61074"} Mar 17 16:15:01 crc kubenswrapper[4767]: I0317 16:15:01.501433 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" podStartSLOduration=1.5014088939999999 podStartE2EDuration="1.501408894s" podCreationTimestamp="2026-03-17 16:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:15:01.493062206 +0000 UTC m=+2292.906378273" watchObservedRunningTime="2026-03-17 16:15:01.501408894 +0000 UTC m=+2292.914724941" Mar 17 16:15:02 crc kubenswrapper[4767]: I0317 16:15:02.502318 4767 generic.go:334] "Generic (PLEG): container finished" podID="7ab4af60-5262-4a85-b749-961ca0b7b162" containerID="881f1867e56ab09c9385353926ad2b79d8af28c1cae6ef12f5934ce58d1608b2" exitCode=0 Mar 17 16:15:02 crc kubenswrapper[4767]: I0317 16:15:02.502976 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" event={"ID":"7ab4af60-5262-4a85-b749-961ca0b7b162","Type":"ContainerDied","Data":"881f1867e56ab09c9385353926ad2b79d8af28c1cae6ef12f5934ce58d1608b2"} Mar 17 16:15:03 crc kubenswrapper[4767]: I0317 16:15:03.524417 4767 generic.go:334] "Generic (PLEG): container finished" podID="9104d786-64f1-41a2-9f89-1863c3101edd" containerID="b0f1a04e8d8770f7683f3c1b53ad891735b9c786f3413f43157c8d9ca9cd82bc" exitCode=0 Mar 17 16:15:03 crc kubenswrapper[4767]: I0317 16:15:03.524474 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" event={"ID":"9104d786-64f1-41a2-9f89-1863c3101edd","Type":"ContainerDied","Data":"b0f1a04e8d8770f7683f3c1b53ad891735b9c786f3413f43157c8d9ca9cd82bc"} Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.080360 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.167209 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.167356 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.256883 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qhcv\" (UniqueName: \"kubernetes.io/projected/7ab4af60-5262-4a85-b749-961ca0b7b162-kube-api-access-4qhcv\") pod \"7ab4af60-5262-4a85-b749-961ca0b7b162\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.257106 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ab4af60-5262-4a85-b749-961ca0b7b162-config-volume\") pod \"7ab4af60-5262-4a85-b749-961ca0b7b162\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.257267 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ab4af60-5262-4a85-b749-961ca0b7b162-secret-volume\") pod \"7ab4af60-5262-4a85-b749-961ca0b7b162\" (UID: \"7ab4af60-5262-4a85-b749-961ca0b7b162\") " Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.258054 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ab4af60-5262-4a85-b749-961ca0b7b162-config-volume" (OuterVolumeSpecName: "config-volume") pod "7ab4af60-5262-4a85-b749-961ca0b7b162" (UID: "7ab4af60-5262-4a85-b749-961ca0b7b162"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.263465 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ab4af60-5262-4a85-b749-961ca0b7b162-kube-api-access-4qhcv" (OuterVolumeSpecName: "kube-api-access-4qhcv") pod "7ab4af60-5262-4a85-b749-961ca0b7b162" (UID: "7ab4af60-5262-4a85-b749-961ca0b7b162"). InnerVolumeSpecName "kube-api-access-4qhcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.263601 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ab4af60-5262-4a85-b749-961ca0b7b162-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7ab4af60-5262-4a85-b749-961ca0b7b162" (UID: "7ab4af60-5262-4a85-b749-961ca0b7b162"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.360669 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qhcv\" (UniqueName: \"kubernetes.io/projected/7ab4af60-5262-4a85-b749-961ca0b7b162-kube-api-access-4qhcv\") on node \"crc\" DevicePath \"\"" Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.360702 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7ab4af60-5262-4a85-b749-961ca0b7b162-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.360711 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7ab4af60-5262-4a85-b749-961ca0b7b162-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.539461 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" event={"ID":"7ab4af60-5262-4a85-b749-961ca0b7b162","Type":"ContainerDied","Data":"4572ae94a921ca68520ee0446c557dc08989b5109e94b1562b33c05ba8d61074"} Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.539521 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4572ae94a921ca68520ee0446c557dc08989b5109e94b1562b33c05ba8d61074" Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.539516 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl" Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.591580 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f"] Mar 17 16:15:04 crc kubenswrapper[4767]: I0317 16:15:04.608426 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562690-fnx5f"] Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.035261 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.188331 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp7hg\" (UniqueName: \"kubernetes.io/projected/9104d786-64f1-41a2-9f89-1863c3101edd-kube-api-access-qp7hg\") pod \"9104d786-64f1-41a2-9f89-1863c3101edd\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.188769 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-inventory\") pod \"9104d786-64f1-41a2-9f89-1863c3101edd\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.188813 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-ssh-key-openstack-edpm-ipam\") pod \"9104d786-64f1-41a2-9f89-1863c3101edd\" (UID: \"9104d786-64f1-41a2-9f89-1863c3101edd\") " Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.197905 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9104d786-64f1-41a2-9f89-1863c3101edd-kube-api-access-qp7hg" (OuterVolumeSpecName: "kube-api-access-qp7hg") pod "9104d786-64f1-41a2-9f89-1863c3101edd" (UID: "9104d786-64f1-41a2-9f89-1863c3101edd"). InnerVolumeSpecName "kube-api-access-qp7hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.233729 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-inventory" (OuterVolumeSpecName: "inventory") pod "9104d786-64f1-41a2-9f89-1863c3101edd" (UID: "9104d786-64f1-41a2-9f89-1863c3101edd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.236065 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9104d786-64f1-41a2-9f89-1863c3101edd" (UID: "9104d786-64f1-41a2-9f89-1863c3101edd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.293714 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.293777 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9104d786-64f1-41a2-9f89-1863c3101edd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.293796 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp7hg\" (UniqueName: \"kubernetes.io/projected/9104d786-64f1-41a2-9f89-1863c3101edd-kube-api-access-qp7hg\") on node \"crc\" DevicePath \"\"" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.376543 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc" path="/var/lib/kubelet/pods/2be6e42f-6e0a-4507-aa9d-6b0e1c297fdc/volumes" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.553363 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" event={"ID":"9104d786-64f1-41a2-9f89-1863c3101edd","Type":"ContainerDied","Data":"288adf1682fa197151c2375f956d865ebe6e34f762776b6bb67eb900586165fb"} Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.553410 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="288adf1682fa197151c2375f956d865ebe6e34f762776b6bb67eb900586165fb" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.553502 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sxqlp" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.644051 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k"] Mar 17 16:15:05 crc kubenswrapper[4767]: E0317 16:15:05.644831 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ab4af60-5262-4a85-b749-961ca0b7b162" containerName="collect-profiles" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.644862 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ab4af60-5262-4a85-b749-961ca0b7b162" containerName="collect-profiles" Mar 17 16:15:05 crc kubenswrapper[4767]: E0317 16:15:05.644935 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9104d786-64f1-41a2-9f89-1863c3101edd" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.644946 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="9104d786-64f1-41a2-9f89-1863c3101edd" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.645441 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ab4af60-5262-4a85-b749-961ca0b7b162" containerName="collect-profiles" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.645510 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="9104d786-64f1-41a2-9f89-1863c3101edd" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.647017 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.652760 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.653164 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.653345 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.653467 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.668355 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k"] Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.809809 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl29f\" (UniqueName: \"kubernetes.io/projected/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-kube-api-access-dl29f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-b255k\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.810467 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-b255k\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.810609 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-b255k\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.913133 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-b255k\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.913337 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-b255k\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.913460 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl29f\" (UniqueName: \"kubernetes.io/projected/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-kube-api-access-dl29f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-b255k\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.925036 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-b255k\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.925114 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-b255k\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.942062 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl29f\" (UniqueName: \"kubernetes.io/projected/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-kube-api-access-dl29f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-b255k\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:15:05 crc kubenswrapper[4767]: I0317 16:15:05.987269 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:15:06 crc kubenswrapper[4767]: I0317 16:15:06.652160 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k"] Mar 17 16:15:07 crc kubenswrapper[4767]: I0317 16:15:07.581444 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" event={"ID":"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20","Type":"ContainerStarted","Data":"4f7c36a7d23e6bdf09f81001b51cc965622d287430f8811306385a5a015e23ce"} Mar 17 16:15:09 crc kubenswrapper[4767]: I0317 16:15:09.611408 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" event={"ID":"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20","Type":"ContainerStarted","Data":"5288b577a17a27cfe4635284a2b36c5af4c496f37ffaa971993736a083447cab"} Mar 17 16:15:09 crc kubenswrapper[4767]: I0317 16:15:09.636661 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" podStartSLOduration=3.569392952 podStartE2EDuration="4.63663693s" podCreationTimestamp="2026-03-17 16:15:05 +0000 UTC" firstStartedPulling="2026-03-17 16:15:06.660646025 +0000 UTC m=+2298.073962072" lastFinishedPulling="2026-03-17 16:15:07.727890003 +0000 UTC m=+2299.141206050" observedRunningTime="2026-03-17 16:15:09.633051407 +0000 UTC m=+2301.046367484" watchObservedRunningTime="2026-03-17 16:15:09.63663693 +0000 UTC m=+2301.049952997" Mar 17 16:15:18 crc kubenswrapper[4767]: I0317 16:15:18.899924 4767 scope.go:117] "RemoveContainer" containerID="e96194c126e9320fbc653d93eb1880202233241335ea78ae4ab4b7862abce819" Mar 17 16:15:18 crc kubenswrapper[4767]: I0317 16:15:18.938300 4767 scope.go:117] "RemoveContainer" containerID="0729e2cd2713c04092f2dd54f9c0acd5be84ae27edf51f2f8ef07b34198bf8bd" Mar 17 16:15:18 crc kubenswrapper[4767]: I0317 16:15:18.991749 4767 scope.go:117] "RemoveContainer" containerID="ece2197718375ad2e2b7f83d4da1945e4eacd599465e12964e926042e8ab2db8" Mar 17 16:15:19 crc kubenswrapper[4767]: I0317 16:15:19.066160 4767 scope.go:117] "RemoveContainer" containerID="30521a8308aa007e0c56096559f18a5842284b1cb62c937699d69295f3e22e08" Mar 17 16:15:19 crc kubenswrapper[4767]: I0317 16:15:19.124367 4767 scope.go:117] "RemoveContainer" containerID="95da0a180a4468719b26f8b482508ee25c55c837077e708b354b1bcc56c9edb7" Mar 17 16:15:19 crc kubenswrapper[4767]: I0317 16:15:19.186304 4767 scope.go:117] "RemoveContainer" containerID="71d626288ff5a1872ae4c9abd0de3e7bcf108b51731d0afcaf32bbbe4ab42d94" Mar 17 16:15:19 crc kubenswrapper[4767]: I0317 16:15:19.210990 4767 scope.go:117] "RemoveContainer" containerID="234a094b33ceb43d9c791e5f7c2816e36f2e876e7a9ff2be4eea8553cb6be82c" Mar 17 16:15:19 crc kubenswrapper[4767]: I0317 16:15:19.280008 4767 scope.go:117] "RemoveContainer" containerID="cb8c86b937458536705cdf5e032ded29e32be4bf131efaabc33199b39b638764" Mar 17 16:15:19 crc kubenswrapper[4767]: I0317 16:15:19.325839 4767 scope.go:117] "RemoveContainer" containerID="72df4826ccc3d3f9a8847830042fbbc37df929e953b73a207af394e476ca5a43" Mar 17 16:15:34 crc kubenswrapper[4767]: I0317 16:15:34.167009 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:15:34 crc kubenswrapper[4767]: I0317 16:15:34.167650 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.574387 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hqqbp"] Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.580160 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.609993 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hqqbp"] Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.656749 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/349fc544-d564-4b99-bac0-bfba49fe72db-catalog-content\") pod \"certified-operators-hqqbp\" (UID: \"349fc544-d564-4b99-bac0-bfba49fe72db\") " pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.656817 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/349fc544-d564-4b99-bac0-bfba49fe72db-utilities\") pod \"certified-operators-hqqbp\" (UID: \"349fc544-d564-4b99-bac0-bfba49fe72db\") " pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.656879 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqkrj\" (UniqueName: \"kubernetes.io/projected/349fc544-d564-4b99-bac0-bfba49fe72db-kube-api-access-cqkrj\") pod \"certified-operators-hqqbp\" (UID: \"349fc544-d564-4b99-bac0-bfba49fe72db\") " pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.760306 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/349fc544-d564-4b99-bac0-bfba49fe72db-catalog-content\") pod \"certified-operators-hqqbp\" (UID: \"349fc544-d564-4b99-bac0-bfba49fe72db\") " pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.760391 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/349fc544-d564-4b99-bac0-bfba49fe72db-utilities\") pod \"certified-operators-hqqbp\" (UID: \"349fc544-d564-4b99-bac0-bfba49fe72db\") " pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.760447 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqkrj\" (UniqueName: \"kubernetes.io/projected/349fc544-d564-4b99-bac0-bfba49fe72db-kube-api-access-cqkrj\") pod \"certified-operators-hqqbp\" (UID: \"349fc544-d564-4b99-bac0-bfba49fe72db\") " pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.760905 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/349fc544-d564-4b99-bac0-bfba49fe72db-catalog-content\") pod \"certified-operators-hqqbp\" (UID: \"349fc544-d564-4b99-bac0-bfba49fe72db\") " pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.761095 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/349fc544-d564-4b99-bac0-bfba49fe72db-utilities\") pod \"certified-operators-hqqbp\" (UID: \"349fc544-d564-4b99-bac0-bfba49fe72db\") " pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.794016 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqkrj\" (UniqueName: \"kubernetes.io/projected/349fc544-d564-4b99-bac0-bfba49fe72db-kube-api-access-cqkrj\") pod \"certified-operators-hqqbp\" (UID: \"349fc544-d564-4b99-bac0-bfba49fe72db\") " pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:15:47 crc kubenswrapper[4767]: I0317 16:15:47.928597 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:15:48 crc kubenswrapper[4767]: I0317 16:15:48.484607 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hqqbp"] Mar 17 16:15:49 crc kubenswrapper[4767]: I0317 16:15:49.164076 4767 generic.go:334] "Generic (PLEG): container finished" podID="349fc544-d564-4b99-bac0-bfba49fe72db" containerID="989aac9e41ca286aee18c919d54252af21106727e1285a4825f5c55c5ee6cfcf" exitCode=0 Mar 17 16:15:49 crc kubenswrapper[4767]: I0317 16:15:49.164192 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqqbp" event={"ID":"349fc544-d564-4b99-bac0-bfba49fe72db","Type":"ContainerDied","Data":"989aac9e41ca286aee18c919d54252af21106727e1285a4825f5c55c5ee6cfcf"} Mar 17 16:15:49 crc kubenswrapper[4767]: I0317 16:15:49.164667 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqqbp" event={"ID":"349fc544-d564-4b99-bac0-bfba49fe72db","Type":"ContainerStarted","Data":"fa4bbd7f76ab0f9bcaabd83a80899005a09c6b84e682ecc60cfff90db9e0eaf4"} Mar 17 16:15:57 crc kubenswrapper[4767]: I0317 16:15:57.051117 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dsn5f"] Mar 17 16:15:57 crc kubenswrapper[4767]: I0317 16:15:57.064615 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dsn5f"] Mar 17 16:15:57 crc kubenswrapper[4767]: I0317 16:15:57.293772 4767 generic.go:334] "Generic (PLEG): container finished" podID="349fc544-d564-4b99-bac0-bfba49fe72db" containerID="0d72d901bff84ffe50c7cd8549550b48df5ebcd5724eb3d598e9b105a7a590e1" exitCode=0 Mar 17 16:15:57 crc kubenswrapper[4767]: I0317 16:15:57.293871 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqqbp" event={"ID":"349fc544-d564-4b99-bac0-bfba49fe72db","Type":"ContainerDied","Data":"0d72d901bff84ffe50c7cd8549550b48df5ebcd5724eb3d598e9b105a7a590e1"} Mar 17 16:15:57 crc kubenswrapper[4767]: I0317 16:15:57.367567 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dba94162-c0cc-46e7-a4ca-45836a003a08" path="/var/lib/kubelet/pods/dba94162-c0cc-46e7-a4ca-45836a003a08/volumes" Mar 17 16:15:58 crc kubenswrapper[4767]: I0317 16:15:58.312041 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqqbp" event={"ID":"349fc544-d564-4b99-bac0-bfba49fe72db","Type":"ContainerStarted","Data":"1f8ea0bd67ee0d5b1ba6dcc640cd65e981a5569550cb3d1bfdd68586e9fddfce"} Mar 17 16:15:58 crc kubenswrapper[4767]: I0317 16:15:58.340689 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hqqbp" podStartSLOduration=2.7216396720000002 podStartE2EDuration="11.34065513s" podCreationTimestamp="2026-03-17 16:15:47 +0000 UTC" firstStartedPulling="2026-03-17 16:15:49.166621067 +0000 UTC m=+2340.579937114" lastFinishedPulling="2026-03-17 16:15:57.785636525 +0000 UTC m=+2349.198952572" observedRunningTime="2026-03-17 16:15:58.33154895 +0000 UTC m=+2349.744865007" watchObservedRunningTime="2026-03-17 16:15:58.34065513 +0000 UTC m=+2349.753971177" Mar 17 16:16:00 crc kubenswrapper[4767]: I0317 16:16:00.155915 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562736-gnl2g"] Mar 17 16:16:00 crc kubenswrapper[4767]: I0317 16:16:00.158571 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562736-gnl2g" Mar 17 16:16:00 crc kubenswrapper[4767]: I0317 16:16:00.161607 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:16:00 crc kubenswrapper[4767]: I0317 16:16:00.162378 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:16:00 crc kubenswrapper[4767]: I0317 16:16:00.162396 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:16:00 crc kubenswrapper[4767]: I0317 16:16:00.181934 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562736-gnl2g"] Mar 17 16:16:00 crc kubenswrapper[4767]: I0317 16:16:00.316337 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlprl\" (UniqueName: \"kubernetes.io/projected/3c54e6dd-f156-4dcb-a753-a3a802735571-kube-api-access-rlprl\") pod \"auto-csr-approver-29562736-gnl2g\" (UID: \"3c54e6dd-f156-4dcb-a753-a3a802735571\") " pod="openshift-infra/auto-csr-approver-29562736-gnl2g" Mar 17 16:16:00 crc kubenswrapper[4767]: I0317 16:16:00.420240 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlprl\" (UniqueName: \"kubernetes.io/projected/3c54e6dd-f156-4dcb-a753-a3a802735571-kube-api-access-rlprl\") pod \"auto-csr-approver-29562736-gnl2g\" (UID: \"3c54e6dd-f156-4dcb-a753-a3a802735571\") " pod="openshift-infra/auto-csr-approver-29562736-gnl2g" Mar 17 16:16:00 crc kubenswrapper[4767]: I0317 16:16:00.446727 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlprl\" (UniqueName: \"kubernetes.io/projected/3c54e6dd-f156-4dcb-a753-a3a802735571-kube-api-access-rlprl\") pod \"auto-csr-approver-29562736-gnl2g\" (UID: \"3c54e6dd-f156-4dcb-a753-a3a802735571\") " pod="openshift-infra/auto-csr-approver-29562736-gnl2g" Mar 17 16:16:00 crc kubenswrapper[4767]: I0317 16:16:00.491330 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562736-gnl2g" Mar 17 16:16:01 crc kubenswrapper[4767]: I0317 16:16:01.034980 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562736-gnl2g"] Mar 17 16:16:01 crc kubenswrapper[4767]: I0317 16:16:01.352039 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562736-gnl2g" event={"ID":"3c54e6dd-f156-4dcb-a753-a3a802735571","Type":"ContainerStarted","Data":"90f36f12d3f31f5776da2ccdbd190c48c88ce363e29fc52b3adc8dd1d3bae534"} Mar 17 16:16:03 crc kubenswrapper[4767]: I0317 16:16:03.386863 4767 generic.go:334] "Generic (PLEG): container finished" podID="3c54e6dd-f156-4dcb-a753-a3a802735571" containerID="2c0ea4042d1e820525c788a4678a5648dde8618947b37195993f7797731cd96e" exitCode=0 Mar 17 16:16:03 crc kubenswrapper[4767]: I0317 16:16:03.387542 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562736-gnl2g" event={"ID":"3c54e6dd-f156-4dcb-a753-a3a802735571","Type":"ContainerDied","Data":"2c0ea4042d1e820525c788a4678a5648dde8618947b37195993f7797731cd96e"} Mar 17 16:16:04 crc kubenswrapper[4767]: I0317 16:16:04.166646 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:16:04 crc kubenswrapper[4767]: I0317 16:16:04.166734 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:16:04 crc kubenswrapper[4767]: I0317 16:16:04.166805 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 16:16:04 crc kubenswrapper[4767]: I0317 16:16:04.168204 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 16:16:04 crc kubenswrapper[4767]: I0317 16:16:04.168283 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" gracePeriod=600 Mar 17 16:16:04 crc kubenswrapper[4767]: E0317 16:16:04.300193 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:16:04 crc kubenswrapper[4767]: I0317 16:16:04.406389 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" exitCode=0 Mar 17 16:16:04 crc kubenswrapper[4767]: I0317 16:16:04.406479 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36"} Mar 17 16:16:04 crc kubenswrapper[4767]: I0317 16:16:04.406534 4767 scope.go:117] "RemoveContainer" containerID="66ab96d70d85b8ad4af97b2cc9506406fc51b53c84a92e8d73017d6fc8cc0c7f" Mar 17 16:16:04 crc kubenswrapper[4767]: I0317 16:16:04.409189 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:16:04 crc kubenswrapper[4767]: E0317 16:16:04.411798 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:16:04 crc kubenswrapper[4767]: I0317 16:16:04.911715 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562736-gnl2g" Mar 17 16:16:05 crc kubenswrapper[4767]: I0317 16:16:05.013696 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlprl\" (UniqueName: \"kubernetes.io/projected/3c54e6dd-f156-4dcb-a753-a3a802735571-kube-api-access-rlprl\") pod \"3c54e6dd-f156-4dcb-a753-a3a802735571\" (UID: \"3c54e6dd-f156-4dcb-a753-a3a802735571\") " Mar 17 16:16:05 crc kubenswrapper[4767]: I0317 16:16:05.023071 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c54e6dd-f156-4dcb-a753-a3a802735571-kube-api-access-rlprl" (OuterVolumeSpecName: "kube-api-access-rlprl") pod "3c54e6dd-f156-4dcb-a753-a3a802735571" (UID: "3c54e6dd-f156-4dcb-a753-a3a802735571"). InnerVolumeSpecName "kube-api-access-rlprl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:16:05 crc kubenswrapper[4767]: I0317 16:16:05.116596 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlprl\" (UniqueName: \"kubernetes.io/projected/3c54e6dd-f156-4dcb-a753-a3a802735571-kube-api-access-rlprl\") on node \"crc\" DevicePath \"\"" Mar 17 16:16:05 crc kubenswrapper[4767]: I0317 16:16:05.431627 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562736-gnl2g" event={"ID":"3c54e6dd-f156-4dcb-a753-a3a802735571","Type":"ContainerDied","Data":"90f36f12d3f31f5776da2ccdbd190c48c88ce363e29fc52b3adc8dd1d3bae534"} Mar 17 16:16:05 crc kubenswrapper[4767]: I0317 16:16:05.431886 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90f36f12d3f31f5776da2ccdbd190c48c88ce363e29fc52b3adc8dd1d3bae534" Mar 17 16:16:05 crc kubenswrapper[4767]: I0317 16:16:05.431685 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562736-gnl2g" Mar 17 16:16:05 crc kubenswrapper[4767]: I0317 16:16:05.982850 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562730-nw2dp"] Mar 17 16:16:05 crc kubenswrapper[4767]: I0317 16:16:05.994978 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562730-nw2dp"] Mar 17 16:16:07 crc kubenswrapper[4767]: I0317 16:16:07.371688 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49eb134e-cc05-459a-959b-e37e432d642a" path="/var/lib/kubelet/pods/49eb134e-cc05-459a-959b-e37e432d642a/volumes" Mar 17 16:16:07 crc kubenswrapper[4767]: I0317 16:16:07.929095 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:16:07 crc kubenswrapper[4767]: I0317 16:16:07.929508 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:16:07 crc kubenswrapper[4767]: I0317 16:16:07.991436 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:16:08 crc kubenswrapper[4767]: I0317 16:16:08.522470 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hqqbp" Mar 17 16:16:08 crc kubenswrapper[4767]: I0317 16:16:08.634225 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hqqbp"] Mar 17 16:16:08 crc kubenswrapper[4767]: I0317 16:16:08.697593 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wnf5c"] Mar 17 16:16:08 crc kubenswrapper[4767]: I0317 16:16:08.698462 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wnf5c" podUID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" containerName="registry-server" containerID="cri-o://f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512" gracePeriod=2 Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.259193 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.309461 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-catalog-content\") pod \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.309706 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnpt5\" (UniqueName: \"kubernetes.io/projected/25c1b139-cd56-4128-ac94-9c0b537f0c0d-kube-api-access-hnpt5\") pod \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.309738 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-utilities\") pod \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\" (UID: \"25c1b139-cd56-4128-ac94-9c0b537f0c0d\") " Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.311438 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-utilities" (OuterVolumeSpecName: "utilities") pod "25c1b139-cd56-4128-ac94-9c0b537f0c0d" (UID: "25c1b139-cd56-4128-ac94-9c0b537f0c0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.322595 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25c1b139-cd56-4128-ac94-9c0b537f0c0d-kube-api-access-hnpt5" (OuterVolumeSpecName: "kube-api-access-hnpt5") pod "25c1b139-cd56-4128-ac94-9c0b537f0c0d" (UID: "25c1b139-cd56-4128-ac94-9c0b537f0c0d"). InnerVolumeSpecName "kube-api-access-hnpt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.414165 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnpt5\" (UniqueName: \"kubernetes.io/projected/25c1b139-cd56-4128-ac94-9c0b537f0c0d-kube-api-access-hnpt5\") on node \"crc\" DevicePath \"\"" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.414267 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.426132 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25c1b139-cd56-4128-ac94-9c0b537f0c0d" (UID: "25c1b139-cd56-4128-ac94-9c0b537f0c0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.488935 4767 generic.go:334] "Generic (PLEG): container finished" podID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" containerID="f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512" exitCode=0 Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.490099 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnf5c" event={"ID":"25c1b139-cd56-4128-ac94-9c0b537f0c0d","Type":"ContainerDied","Data":"f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512"} Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.490213 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnf5c" event={"ID":"25c1b139-cd56-4128-ac94-9c0b537f0c0d","Type":"ContainerDied","Data":"36e4d9531b81bcc28adf6e2021ac0ebb1eb4dcf767078cc24e2a922c0c3afabf"} Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.490242 4767 scope.go:117] "RemoveContainer" containerID="f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.490132 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnf5c" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.519945 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c1b139-cd56-4128-ac94-9c0b537f0c0d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.528585 4767 scope.go:117] "RemoveContainer" containerID="7531c28fdd615c44e17b1414549a58e4c9498789ecb3364760c2db5569c5f29d" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.567621 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wnf5c"] Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.578148 4767 scope.go:117] "RemoveContainer" containerID="fa047f500d41d4af3255fb1beb5ea59580f132d6384a445aad7338207becee0c" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.590752 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wnf5c"] Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.634323 4767 scope.go:117] "RemoveContainer" containerID="f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512" Mar 17 16:16:09 crc kubenswrapper[4767]: E0317 16:16:09.635519 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512\": container with ID starting with f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512 not found: ID does not exist" containerID="f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.635587 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512"} err="failed to get container status \"f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512\": rpc error: code = NotFound desc = could not find container \"f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512\": container with ID starting with f7d176a9545099eb7f8e2f8b14499baf496eede5ba6e11e939c91e4409dd6512 not found: ID does not exist" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.635630 4767 scope.go:117] "RemoveContainer" containerID="7531c28fdd615c44e17b1414549a58e4c9498789ecb3364760c2db5569c5f29d" Mar 17 16:16:09 crc kubenswrapper[4767]: E0317 16:16:09.636060 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7531c28fdd615c44e17b1414549a58e4c9498789ecb3364760c2db5569c5f29d\": container with ID starting with 7531c28fdd615c44e17b1414549a58e4c9498789ecb3364760c2db5569c5f29d not found: ID does not exist" containerID="7531c28fdd615c44e17b1414549a58e4c9498789ecb3364760c2db5569c5f29d" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.636116 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7531c28fdd615c44e17b1414549a58e4c9498789ecb3364760c2db5569c5f29d"} err="failed to get container status \"7531c28fdd615c44e17b1414549a58e4c9498789ecb3364760c2db5569c5f29d\": rpc error: code = NotFound desc = could not find container \"7531c28fdd615c44e17b1414549a58e4c9498789ecb3364760c2db5569c5f29d\": container with ID starting with 7531c28fdd615c44e17b1414549a58e4c9498789ecb3364760c2db5569c5f29d not found: ID does not exist" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.636158 4767 scope.go:117] "RemoveContainer" containerID="fa047f500d41d4af3255fb1beb5ea59580f132d6384a445aad7338207becee0c" Mar 17 16:16:09 crc kubenswrapper[4767]: E0317 16:16:09.636960 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa047f500d41d4af3255fb1beb5ea59580f132d6384a445aad7338207becee0c\": container with ID starting with fa047f500d41d4af3255fb1beb5ea59580f132d6384a445aad7338207becee0c not found: ID does not exist" containerID="fa047f500d41d4af3255fb1beb5ea59580f132d6384a445aad7338207becee0c" Mar 17 16:16:09 crc kubenswrapper[4767]: I0317 16:16:09.637003 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa047f500d41d4af3255fb1beb5ea59580f132d6384a445aad7338207becee0c"} err="failed to get container status \"fa047f500d41d4af3255fb1beb5ea59580f132d6384a445aad7338207becee0c\": rpc error: code = NotFound desc = could not find container \"fa047f500d41d4af3255fb1beb5ea59580f132d6384a445aad7338207becee0c\": container with ID starting with fa047f500d41d4af3255fb1beb5ea59580f132d6384a445aad7338207becee0c not found: ID does not exist" Mar 17 16:16:11 crc kubenswrapper[4767]: I0317 16:16:11.369795 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" path="/var/lib/kubelet/pods/25c1b139-cd56-4128-ac94-9c0b537f0c0d/volumes" Mar 17 16:16:16 crc kubenswrapper[4767]: I0317 16:16:16.355105 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:16:16 crc kubenswrapper[4767]: E0317 16:16:16.356111 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:16:18 crc kubenswrapper[4767]: I0317 16:16:18.054072 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-nlvws"] Mar 17 16:16:18 crc kubenswrapper[4767]: I0317 16:16:18.070037 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-21c1-account-create-update-2g64g"] Mar 17 16:16:18 crc kubenswrapper[4767]: I0317 16:16:18.082616 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-21c1-account-create-update-2g64g"] Mar 17 16:16:18 crc kubenswrapper[4767]: I0317 16:16:18.094006 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-nlvws"] Mar 17 16:16:19 crc kubenswrapper[4767]: I0317 16:16:19.371387 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="271cc3d9-d20c-4708-bd6a-765a8c439cf6" path="/var/lib/kubelet/pods/271cc3d9-d20c-4708-bd6a-765a8c439cf6/volumes" Mar 17 16:16:19 crc kubenswrapper[4767]: I0317 16:16:19.378662 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46218023-25a1-49e3-8358-08445caf00ce" path="/var/lib/kubelet/pods/46218023-25a1-49e3-8358-08445caf00ce/volumes" Mar 17 16:16:19 crc kubenswrapper[4767]: I0317 16:16:19.613662 4767 scope.go:117] "RemoveContainer" containerID="19902911191b266a4117332e9664ffbe0b11a0de4db4a9275ffc9802b5c13c9e" Mar 17 16:16:19 crc kubenswrapper[4767]: I0317 16:16:19.650856 4767 scope.go:117] "RemoveContainer" containerID="f02b1b789c7fab91bfdcb241feccd8e4d944cf6ae3e4f0e3064689dcd7c3742b" Mar 17 16:16:19 crc kubenswrapper[4767]: I0317 16:16:19.758343 4767 scope.go:117] "RemoveContainer" containerID="eea4f2a1ea128dac6e5fa290ffcaf4f7e32b29d8e34f63c19f28c0eebd49d236" Mar 17 16:16:19 crc kubenswrapper[4767]: I0317 16:16:19.792496 4767 scope.go:117] "RemoveContainer" containerID="fb149d46dfe5eb769cecad148cab00998aa2fac8546e22b41cf3012056ebbb54" Mar 17 16:16:21 crc kubenswrapper[4767]: I0317 16:16:21.634063 4767 generic.go:334] "Generic (PLEG): container finished" podID="0f019c8e-b34a-4a66-9387-1bdf6e5d6b20" containerID="5288b577a17a27cfe4635284a2b36c5af4c496f37ffaa971993736a083447cab" exitCode=0 Mar 17 16:16:21 crc kubenswrapper[4767]: I0317 16:16:21.634148 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" event={"ID":"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20","Type":"ContainerDied","Data":"5288b577a17a27cfe4635284a2b36c5af4c496f37ffaa971993736a083447cab"} Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.135499 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.246549 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-inventory\") pod \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.246612 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-ssh-key-openstack-edpm-ipam\") pod \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.246648 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl29f\" (UniqueName: \"kubernetes.io/projected/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-kube-api-access-dl29f\") pod \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\" (UID: \"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20\") " Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.253478 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-kube-api-access-dl29f" (OuterVolumeSpecName: "kube-api-access-dl29f") pod "0f019c8e-b34a-4a66-9387-1bdf6e5d6b20" (UID: "0f019c8e-b34a-4a66-9387-1bdf6e5d6b20"). InnerVolumeSpecName "kube-api-access-dl29f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.280864 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0f019c8e-b34a-4a66-9387-1bdf6e5d6b20" (UID: "0f019c8e-b34a-4a66-9387-1bdf6e5d6b20"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.290090 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-inventory" (OuterVolumeSpecName: "inventory") pod "0f019c8e-b34a-4a66-9387-1bdf6e5d6b20" (UID: "0f019c8e-b34a-4a66-9387-1bdf6e5d6b20"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.349670 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.350058 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.350073 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl29f\" (UniqueName: \"kubernetes.io/projected/0f019c8e-b34a-4a66-9387-1bdf6e5d6b20-kube-api-access-dl29f\") on node \"crc\" DevicePath \"\"" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.671461 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" event={"ID":"0f019c8e-b34a-4a66-9387-1bdf6e5d6b20","Type":"ContainerDied","Data":"4f7c36a7d23e6bdf09f81001b51cc965622d287430f8811306385a5a015e23ce"} Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.671510 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f7c36a7d23e6bdf09f81001b51cc965622d287430f8811306385a5a015e23ce" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.671667 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-b255k" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.778580 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v"] Mar 17 16:16:23 crc kubenswrapper[4767]: E0317 16:16:23.780518 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" containerName="extract-content" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.780653 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" containerName="extract-content" Mar 17 16:16:23 crc kubenswrapper[4767]: E0317 16:16:23.780986 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" containerName="registry-server" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.781072 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" containerName="registry-server" Mar 17 16:16:23 crc kubenswrapper[4767]: E0317 16:16:23.781183 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c54e6dd-f156-4dcb-a753-a3a802735571" containerName="oc" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.781270 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c54e6dd-f156-4dcb-a753-a3a802735571" containerName="oc" Mar 17 16:16:23 crc kubenswrapper[4767]: E0317 16:16:23.781397 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f019c8e-b34a-4a66-9387-1bdf6e5d6b20" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.781494 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f019c8e-b34a-4a66-9387-1bdf6e5d6b20" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 17 16:16:23 crc kubenswrapper[4767]: E0317 16:16:23.781613 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" containerName="extract-utilities" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.781706 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" containerName="extract-utilities" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.788391 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f019c8e-b34a-4a66-9387-1bdf6e5d6b20" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.788669 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c54e6dd-f156-4dcb-a753-a3a802735571" containerName="oc" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.788829 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="25c1b139-cd56-4128-ac94-9c0b537f0c0d" containerName="registry-server" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.790011 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.795199 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.795538 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.795700 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.797163 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v"] Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.811334 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.875355 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.875406 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4vkx\" (UniqueName: \"kubernetes.io/projected/da025599-26ff-4572-be75-97d0a3017e90-kube-api-access-q4vkx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.875581 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.978581 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.978636 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4vkx\" (UniqueName: \"kubernetes.io/projected/da025599-26ff-4572-be75-97d0a3017e90-kube-api-access-q4vkx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.978747 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.983094 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.983600 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:23 crc kubenswrapper[4767]: I0317 16:16:23.999820 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4vkx\" (UniqueName: \"kubernetes.io/projected/da025599-26ff-4572-be75-97d0a3017e90-kube-api-access-q4vkx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:24 crc kubenswrapper[4767]: I0317 16:16:24.124913 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:24 crc kubenswrapper[4767]: I0317 16:16:24.686011 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v"] Mar 17 16:16:25 crc kubenswrapper[4767]: I0317 16:16:25.699873 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" event={"ID":"da025599-26ff-4572-be75-97d0a3017e90","Type":"ContainerStarted","Data":"da26f8dbcf015b895fb59e415a01a3cc33ffb8e2a073443c0fef2f1e0caa1732"} Mar 17 16:16:25 crc kubenswrapper[4767]: I0317 16:16:25.700340 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" event={"ID":"da025599-26ff-4572-be75-97d0a3017e90","Type":"ContainerStarted","Data":"e77d10756f30f0ced786a8fdb08d27da1454b205da237d14747890e4e19ee21a"} Mar 17 16:16:28 crc kubenswrapper[4767]: I0317 16:16:28.354976 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:16:28 crc kubenswrapper[4767]: E0317 16:16:28.355988 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:16:30 crc kubenswrapper[4767]: I0317 16:16:30.059850 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" podStartSLOduration=6.462580888 podStartE2EDuration="7.059817275s" podCreationTimestamp="2026-03-17 16:16:23 +0000 UTC" firstStartedPulling="2026-03-17 16:16:24.695823831 +0000 UTC m=+2376.109139878" lastFinishedPulling="2026-03-17 16:16:25.293060198 +0000 UTC m=+2376.706376265" observedRunningTime="2026-03-17 16:16:25.720107446 +0000 UTC m=+2377.133423493" watchObservedRunningTime="2026-03-17 16:16:30.059817275 +0000 UTC m=+2381.473133332" Mar 17 16:16:30 crc kubenswrapper[4767]: I0317 16:16:30.063400 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-p7cb5"] Mar 17 16:16:30 crc kubenswrapper[4767]: I0317 16:16:30.076313 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-p7cb5"] Mar 17 16:16:30 crc kubenswrapper[4767]: I0317 16:16:30.764112 4767 generic.go:334] "Generic (PLEG): container finished" podID="da025599-26ff-4572-be75-97d0a3017e90" containerID="da26f8dbcf015b895fb59e415a01a3cc33ffb8e2a073443c0fef2f1e0caa1732" exitCode=0 Mar 17 16:16:30 crc kubenswrapper[4767]: I0317 16:16:30.764171 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" event={"ID":"da025599-26ff-4572-be75-97d0a3017e90","Type":"ContainerDied","Data":"da26f8dbcf015b895fb59e415a01a3cc33ffb8e2a073443c0fef2f1e0caa1732"} Mar 17 16:16:31 crc kubenswrapper[4767]: I0317 16:16:31.369249 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aaae974-d49d-4980-a540-6bc2ec85a15f" path="/var/lib/kubelet/pods/0aaae974-d49d-4980-a540-6bc2ec85a15f/volumes" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.313995 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.410530 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-ssh-key-openstack-edpm-ipam\") pod \"da025599-26ff-4572-be75-97d0a3017e90\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.410609 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-inventory\") pod \"da025599-26ff-4572-be75-97d0a3017e90\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.410901 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4vkx\" (UniqueName: \"kubernetes.io/projected/da025599-26ff-4572-be75-97d0a3017e90-kube-api-access-q4vkx\") pod \"da025599-26ff-4572-be75-97d0a3017e90\" (UID: \"da025599-26ff-4572-be75-97d0a3017e90\") " Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.417398 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da025599-26ff-4572-be75-97d0a3017e90-kube-api-access-q4vkx" (OuterVolumeSpecName: "kube-api-access-q4vkx") pod "da025599-26ff-4572-be75-97d0a3017e90" (UID: "da025599-26ff-4572-be75-97d0a3017e90"). InnerVolumeSpecName "kube-api-access-q4vkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.454090 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-inventory" (OuterVolumeSpecName: "inventory") pod "da025599-26ff-4572-be75-97d0a3017e90" (UID: "da025599-26ff-4572-be75-97d0a3017e90"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.460000 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "da025599-26ff-4572-be75-97d0a3017e90" (UID: "da025599-26ff-4572-be75-97d0a3017e90"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.514839 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4vkx\" (UniqueName: \"kubernetes.io/projected/da025599-26ff-4572-be75-97d0a3017e90-kube-api-access-q4vkx\") on node \"crc\" DevicePath \"\"" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.514884 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.514895 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da025599-26ff-4572-be75-97d0a3017e90-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.790977 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" event={"ID":"da025599-26ff-4572-be75-97d0a3017e90","Type":"ContainerDied","Data":"e77d10756f30f0ced786a8fdb08d27da1454b205da237d14747890e4e19ee21a"} Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.791522 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e77d10756f30f0ced786a8fdb08d27da1454b205da237d14747890e4e19ee21a" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.791054 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rqk2v" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.882435 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x"] Mar 17 16:16:32 crc kubenswrapper[4767]: E0317 16:16:32.883367 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da025599-26ff-4572-be75-97d0a3017e90" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.883396 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="da025599-26ff-4572-be75-97d0a3017e90" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.883875 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="da025599-26ff-4572-be75-97d0a3017e90" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.885142 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.889207 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.889525 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.894330 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.900125 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:16:32 crc kubenswrapper[4767]: I0317 16:16:32.900223 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x"] Mar 17 16:16:33 crc kubenswrapper[4767]: I0317 16:16:33.037354 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-z2h4x\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:16:33 crc kubenswrapper[4767]: I0317 16:16:33.037879 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67n9z\" (UniqueName: \"kubernetes.io/projected/5eef90f9-51ac-4656-8083-93f8f9032ad9-kube-api-access-67n9z\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-z2h4x\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:16:33 crc kubenswrapper[4767]: I0317 16:16:33.038376 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-z2h4x\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:16:33 crc kubenswrapper[4767]: I0317 16:16:33.141450 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-z2h4x\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:16:33 crc kubenswrapper[4767]: I0317 16:16:33.141702 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67n9z\" (UniqueName: \"kubernetes.io/projected/5eef90f9-51ac-4656-8083-93f8f9032ad9-kube-api-access-67n9z\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-z2h4x\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:16:33 crc kubenswrapper[4767]: I0317 16:16:33.141808 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-z2h4x\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:16:33 crc kubenswrapper[4767]: I0317 16:16:33.147092 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-z2h4x\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:16:33 crc kubenswrapper[4767]: I0317 16:16:33.147131 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-z2h4x\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:16:33 crc kubenswrapper[4767]: I0317 16:16:33.161873 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67n9z\" (UniqueName: \"kubernetes.io/projected/5eef90f9-51ac-4656-8083-93f8f9032ad9-kube-api-access-67n9z\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-z2h4x\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:16:33 crc kubenswrapper[4767]: I0317 16:16:33.208855 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:16:33 crc kubenswrapper[4767]: I0317 16:16:33.808885 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x"] Mar 17 16:16:34 crc kubenswrapper[4767]: I0317 16:16:34.042761 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9fv46"] Mar 17 16:16:34 crc kubenswrapper[4767]: I0317 16:16:34.057270 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9fv46"] Mar 17 16:16:34 crc kubenswrapper[4767]: I0317 16:16:34.846513 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" event={"ID":"5eef90f9-51ac-4656-8083-93f8f9032ad9","Type":"ContainerStarted","Data":"dc5a0f4c7906247af6fa57f140ce380f8ece12056653ffbd6775f5a848119c82"} Mar 17 16:16:35 crc kubenswrapper[4767]: I0317 16:16:35.370601 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3d7ba45-9874-4177-81d0-9c8f88280cfe" path="/var/lib/kubelet/pods/d3d7ba45-9874-4177-81d0-9c8f88280cfe/volumes" Mar 17 16:16:35 crc kubenswrapper[4767]: I0317 16:16:35.859959 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" event={"ID":"5eef90f9-51ac-4656-8083-93f8f9032ad9","Type":"ContainerStarted","Data":"41e48908e065a3026124376bd29624351145b00e64a6aed178bdd753c45aefb5"} Mar 17 16:16:35 crc kubenswrapper[4767]: I0317 16:16:35.893132 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" podStartSLOduration=3.405721115 podStartE2EDuration="3.893104082s" podCreationTimestamp="2026-03-17 16:16:32 +0000 UTC" firstStartedPulling="2026-03-17 16:16:33.819722215 +0000 UTC m=+2385.233038262" lastFinishedPulling="2026-03-17 16:16:34.307105182 +0000 UTC m=+2385.720421229" observedRunningTime="2026-03-17 16:16:35.887379019 +0000 UTC m=+2387.300695266" watchObservedRunningTime="2026-03-17 16:16:35.893104082 +0000 UTC m=+2387.306420139" Mar 17 16:16:39 crc kubenswrapper[4767]: I0317 16:16:39.363539 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:16:39 crc kubenswrapper[4767]: E0317 16:16:39.364438 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:16:50 crc kubenswrapper[4767]: I0317 16:16:50.355702 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:16:50 crc kubenswrapper[4767]: E0317 16:16:50.356578 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:16:50 crc kubenswrapper[4767]: I0317 16:16:50.938059 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-56879488c7-5jbhc" podUID="65779f05-d74b-49a7-a31a-2de7f9005e59" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 17 16:17:04 crc kubenswrapper[4767]: I0317 16:17:04.355022 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:17:04 crc kubenswrapper[4767]: E0317 16:17:04.355831 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:17:10 crc kubenswrapper[4767]: I0317 16:17:10.278022 4767 generic.go:334] "Generic (PLEG): container finished" podID="5eef90f9-51ac-4656-8083-93f8f9032ad9" containerID="41e48908e065a3026124376bd29624351145b00e64a6aed178bdd753c45aefb5" exitCode=0 Mar 17 16:17:10 crc kubenswrapper[4767]: I0317 16:17:10.278116 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" event={"ID":"5eef90f9-51ac-4656-8083-93f8f9032ad9","Type":"ContainerDied","Data":"41e48908e065a3026124376bd29624351145b00e64a6aed178bdd753c45aefb5"} Mar 17 16:17:11 crc kubenswrapper[4767]: I0317 16:17:11.936607 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.001033 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67n9z\" (UniqueName: \"kubernetes.io/projected/5eef90f9-51ac-4656-8083-93f8f9032ad9-kube-api-access-67n9z\") pod \"5eef90f9-51ac-4656-8083-93f8f9032ad9\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.001246 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-ssh-key-openstack-edpm-ipam\") pod \"5eef90f9-51ac-4656-8083-93f8f9032ad9\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.001354 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-inventory\") pod \"5eef90f9-51ac-4656-8083-93f8f9032ad9\" (UID: \"5eef90f9-51ac-4656-8083-93f8f9032ad9\") " Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.033721 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eef90f9-51ac-4656-8083-93f8f9032ad9-kube-api-access-67n9z" (OuterVolumeSpecName: "kube-api-access-67n9z") pod "5eef90f9-51ac-4656-8083-93f8f9032ad9" (UID: "5eef90f9-51ac-4656-8083-93f8f9032ad9"). InnerVolumeSpecName "kube-api-access-67n9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.155124 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5eef90f9-51ac-4656-8083-93f8f9032ad9" (UID: "5eef90f9-51ac-4656-8083-93f8f9032ad9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.159567 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67n9z\" (UniqueName: \"kubernetes.io/projected/5eef90f9-51ac-4656-8083-93f8f9032ad9-kube-api-access-67n9z\") on node \"crc\" DevicePath \"\"" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.212893 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-inventory" (OuterVolumeSpecName: "inventory") pod "5eef90f9-51ac-4656-8083-93f8f9032ad9" (UID: "5eef90f9-51ac-4656-8083-93f8f9032ad9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.267240 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.267293 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5eef90f9-51ac-4656-8083-93f8f9032ad9-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.306226 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" event={"ID":"5eef90f9-51ac-4656-8083-93f8f9032ad9","Type":"ContainerDied","Data":"dc5a0f4c7906247af6fa57f140ce380f8ece12056653ffbd6775f5a848119c82"} Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.306291 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc5a0f4c7906247af6fa57f140ce380f8ece12056653ffbd6775f5a848119c82" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.306367 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-z2h4x" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.410976 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s"] Mar 17 16:17:12 crc kubenswrapper[4767]: E0317 16:17:12.411814 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eef90f9-51ac-4656-8083-93f8f9032ad9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.411838 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eef90f9-51ac-4656-8083-93f8f9032ad9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.412144 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eef90f9-51ac-4656-8083-93f8f9032ad9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.413485 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.417163 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.417527 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.417700 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.417967 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.423097 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s"] Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.472246 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qj64s\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.472316 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qj64s\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.472987 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb4p7\" (UniqueName: \"kubernetes.io/projected/760ad35b-f4a3-4c24-96b6-5b193caf7dae-kube-api-access-jb4p7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qj64s\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.574600 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb4p7\" (UniqueName: \"kubernetes.io/projected/760ad35b-f4a3-4c24-96b6-5b193caf7dae-kube-api-access-jb4p7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qj64s\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.575116 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qj64s\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.575319 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qj64s\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.581706 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qj64s\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.585090 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qj64s\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.597377 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb4p7\" (UniqueName: \"kubernetes.io/projected/760ad35b-f4a3-4c24-96b6-5b193caf7dae-kube-api-access-jb4p7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-qj64s\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:17:12 crc kubenswrapper[4767]: I0317 16:17:12.735505 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:17:13 crc kubenswrapper[4767]: I0317 16:17:13.343474 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s"] Mar 17 16:17:13 crc kubenswrapper[4767]: I0317 16:17:13.355762 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 16:17:14 crc kubenswrapper[4767]: I0317 16:17:14.333578 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" event={"ID":"760ad35b-f4a3-4c24-96b6-5b193caf7dae","Type":"ContainerStarted","Data":"c55234104d6572536bd7323e1fc0166349d4457fc9789b4bd2ba2607a0bc8c94"} Mar 17 16:17:15 crc kubenswrapper[4767]: I0317 16:17:15.345747 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" event={"ID":"760ad35b-f4a3-4c24-96b6-5b193caf7dae","Type":"ContainerStarted","Data":"90bbfda96ba5bb67893858cdacef9c5978e0aad2ea6ee759007bbd3408556ffd"} Mar 17 16:17:15 crc kubenswrapper[4767]: I0317 16:17:15.378471 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" podStartSLOduration=2.655817038 podStartE2EDuration="3.378445038s" podCreationTimestamp="2026-03-17 16:17:12 +0000 UTC" firstStartedPulling="2026-03-17 16:17:13.355521769 +0000 UTC m=+2424.768837816" lastFinishedPulling="2026-03-17 16:17:14.078149769 +0000 UTC m=+2425.491465816" observedRunningTime="2026-03-17 16:17:15.370047489 +0000 UTC m=+2426.783363556" watchObservedRunningTime="2026-03-17 16:17:15.378445038 +0000 UTC m=+2426.791761085" Mar 17 16:17:17 crc kubenswrapper[4767]: I0317 16:17:17.354815 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:17:17 crc kubenswrapper[4767]: E0317 16:17:17.355521 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:17:18 crc kubenswrapper[4767]: I0317 16:17:18.062423 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-vcxwn"] Mar 17 16:17:18 crc kubenswrapper[4767]: I0317 16:17:18.074580 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-vcxwn"] Mar 17 16:17:19 crc kubenswrapper[4767]: I0317 16:17:19.372249 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e2e4698-f52a-476f-8d61-da9459bb6fe0" path="/var/lib/kubelet/pods/3e2e4698-f52a-476f-8d61-da9459bb6fe0/volumes" Mar 17 16:17:20 crc kubenswrapper[4767]: I0317 16:17:20.204869 4767 scope.go:117] "RemoveContainer" containerID="6de3fce5967ad59a9643a96e5e108e2b64cb299cb49ba3c0becfde2ed5e10f36" Mar 17 16:17:20 crc kubenswrapper[4767]: I0317 16:17:20.264281 4767 scope.go:117] "RemoveContainer" containerID="466e5152b4dae52e57a86166113f0ef3d5745577ca45e1f223ca9a7f0065c3f8" Mar 17 16:17:20 crc kubenswrapper[4767]: I0317 16:17:20.326208 4767 scope.go:117] "RemoveContainer" containerID="0299cbebda8a41b6539886434bc5f0eded000fbd93450631eedae08c5e159aeb" Mar 17 16:17:29 crc kubenswrapper[4767]: I0317 16:17:29.363326 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:17:29 crc kubenswrapper[4767]: E0317 16:17:29.364244 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:17:43 crc kubenswrapper[4767]: I0317 16:17:43.355502 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:17:43 crc kubenswrapper[4767]: E0317 16:17:43.356749 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:17:55 crc kubenswrapper[4767]: I0317 16:17:55.354758 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:17:55 crc kubenswrapper[4767]: E0317 16:17:55.357811 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:18:00 crc kubenswrapper[4767]: I0317 16:18:00.161836 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562738-k77dn"] Mar 17 16:18:00 crc kubenswrapper[4767]: I0317 16:18:00.165216 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562738-k77dn" Mar 17 16:18:00 crc kubenswrapper[4767]: I0317 16:18:00.168831 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:18:00 crc kubenswrapper[4767]: I0317 16:18:00.169426 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:18:00 crc kubenswrapper[4767]: I0317 16:18:00.171447 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:18:00 crc kubenswrapper[4767]: I0317 16:18:00.173244 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562738-k77dn"] Mar 17 16:18:00 crc kubenswrapper[4767]: I0317 16:18:00.211021 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmgl5\" (UniqueName: \"kubernetes.io/projected/8bd93414-e722-485f-a4e5-453ef7d6c4aa-kube-api-access-zmgl5\") pod \"auto-csr-approver-29562738-k77dn\" (UID: \"8bd93414-e722-485f-a4e5-453ef7d6c4aa\") " pod="openshift-infra/auto-csr-approver-29562738-k77dn" Mar 17 16:18:00 crc kubenswrapper[4767]: I0317 16:18:00.313312 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmgl5\" (UniqueName: \"kubernetes.io/projected/8bd93414-e722-485f-a4e5-453ef7d6c4aa-kube-api-access-zmgl5\") pod \"auto-csr-approver-29562738-k77dn\" (UID: \"8bd93414-e722-485f-a4e5-453ef7d6c4aa\") " pod="openshift-infra/auto-csr-approver-29562738-k77dn" Mar 17 16:18:00 crc kubenswrapper[4767]: I0317 16:18:00.338019 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmgl5\" (UniqueName: \"kubernetes.io/projected/8bd93414-e722-485f-a4e5-453ef7d6c4aa-kube-api-access-zmgl5\") pod \"auto-csr-approver-29562738-k77dn\" (UID: \"8bd93414-e722-485f-a4e5-453ef7d6c4aa\") " pod="openshift-infra/auto-csr-approver-29562738-k77dn" Mar 17 16:18:00 crc kubenswrapper[4767]: I0317 16:18:00.491131 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562738-k77dn" Mar 17 16:18:01 crc kubenswrapper[4767]: I0317 16:18:01.059356 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562738-k77dn"] Mar 17 16:18:02 crc kubenswrapper[4767]: I0317 16:18:02.026417 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562738-k77dn" event={"ID":"8bd93414-e722-485f-a4e5-453ef7d6c4aa","Type":"ContainerStarted","Data":"14716ede3f69c7be204b009b7c62a2e1285450dcd5943068d91152894f638b38"} Mar 17 16:18:02 crc kubenswrapper[4767]: I0317 16:18:02.032649 4767 generic.go:334] "Generic (PLEG): container finished" podID="760ad35b-f4a3-4c24-96b6-5b193caf7dae" containerID="90bbfda96ba5bb67893858cdacef9c5978e0aad2ea6ee759007bbd3408556ffd" exitCode=0 Mar 17 16:18:02 crc kubenswrapper[4767]: I0317 16:18:02.032761 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" event={"ID":"760ad35b-f4a3-4c24-96b6-5b193caf7dae","Type":"ContainerDied","Data":"90bbfda96ba5bb67893858cdacef9c5978e0aad2ea6ee759007bbd3408556ffd"} Mar 17 16:18:03 crc kubenswrapper[4767]: I0317 16:18:03.089323 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562738-k77dn" event={"ID":"8bd93414-e722-485f-a4e5-453ef7d6c4aa","Type":"ContainerStarted","Data":"2981b86328461469ddfe474c26d47ddd87adcb49a171fd4a7b97b76b6f085f36"} Mar 17 16:18:03 crc kubenswrapper[4767]: I0317 16:18:03.132428 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562738-k77dn" podStartSLOduration=2.298266265 podStartE2EDuration="3.132390416s" podCreationTimestamp="2026-03-17 16:18:00 +0000 UTC" firstStartedPulling="2026-03-17 16:18:01.069903245 +0000 UTC m=+2472.483219292" lastFinishedPulling="2026-03-17 16:18:01.904027396 +0000 UTC m=+2473.317343443" observedRunningTime="2026-03-17 16:18:03.116446135 +0000 UTC m=+2474.529762202" watchObservedRunningTime="2026-03-17 16:18:03.132390416 +0000 UTC m=+2474.545706463" Mar 17 16:18:03 crc kubenswrapper[4767]: I0317 16:18:03.843894 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:18:03 crc kubenswrapper[4767]: I0317 16:18:03.902320 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jb4p7\" (UniqueName: \"kubernetes.io/projected/760ad35b-f4a3-4c24-96b6-5b193caf7dae-kube-api-access-jb4p7\") pod \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " Mar 17 16:18:03 crc kubenswrapper[4767]: I0317 16:18:03.902613 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-ssh-key-openstack-edpm-ipam\") pod \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " Mar 17 16:18:03 crc kubenswrapper[4767]: I0317 16:18:03.902688 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-inventory\") pod \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\" (UID: \"760ad35b-f4a3-4c24-96b6-5b193caf7dae\") " Mar 17 16:18:03 crc kubenswrapper[4767]: I0317 16:18:03.908804 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/760ad35b-f4a3-4c24-96b6-5b193caf7dae-kube-api-access-jb4p7" (OuterVolumeSpecName: "kube-api-access-jb4p7") pod "760ad35b-f4a3-4c24-96b6-5b193caf7dae" (UID: "760ad35b-f4a3-4c24-96b6-5b193caf7dae"). InnerVolumeSpecName "kube-api-access-jb4p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:18:03 crc kubenswrapper[4767]: I0317 16:18:03.955962 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "760ad35b-f4a3-4c24-96b6-5b193caf7dae" (UID: "760ad35b-f4a3-4c24-96b6-5b193caf7dae"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:18:03 crc kubenswrapper[4767]: I0317 16:18:03.967643 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-inventory" (OuterVolumeSpecName: "inventory") pod "760ad35b-f4a3-4c24-96b6-5b193caf7dae" (UID: "760ad35b-f4a3-4c24-96b6-5b193caf7dae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.006571 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.006908 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jb4p7\" (UniqueName: \"kubernetes.io/projected/760ad35b-f4a3-4c24-96b6-5b193caf7dae-kube-api-access-jb4p7\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.006920 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/760ad35b-f4a3-4c24-96b6-5b193caf7dae-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.105986 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" event={"ID":"760ad35b-f4a3-4c24-96b6-5b193caf7dae","Type":"ContainerDied","Data":"c55234104d6572536bd7323e1fc0166349d4457fc9789b4bd2ba2607a0bc8c94"} Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.106032 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-qj64s" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.106073 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c55234104d6572536bd7323e1fc0166349d4457fc9789b4bd2ba2607a0bc8c94" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.112227 4767 generic.go:334] "Generic (PLEG): container finished" podID="8bd93414-e722-485f-a4e5-453ef7d6c4aa" containerID="2981b86328461469ddfe474c26d47ddd87adcb49a171fd4a7b97b76b6f085f36" exitCode=0 Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.112293 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562738-k77dn" event={"ID":"8bd93414-e722-485f-a4e5-453ef7d6c4aa","Type":"ContainerDied","Data":"2981b86328461469ddfe474c26d47ddd87adcb49a171fd4a7b97b76b6f085f36"} Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.175315 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ltqf4"] Mar 17 16:18:04 crc kubenswrapper[4767]: E0317 16:18:04.176497 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="760ad35b-f4a3-4c24-96b6-5b193caf7dae" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.176522 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="760ad35b-f4a3-4c24-96b6-5b193caf7dae" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.176817 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="760ad35b-f4a3-4c24-96b6-5b193caf7dae" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.188878 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.189558 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ltqf4"] Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.193132 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.193190 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.193207 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.193572 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.212515 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffqr8\" (UniqueName: \"kubernetes.io/projected/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-kube-api-access-ffqr8\") pod \"ssh-known-hosts-edpm-deployment-ltqf4\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.212975 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ltqf4\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.213437 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ltqf4\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.333932 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ltqf4\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.334164 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ltqf4\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.334276 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffqr8\" (UniqueName: \"kubernetes.io/projected/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-kube-api-access-ffqr8\") pod \"ssh-known-hosts-edpm-deployment-ltqf4\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.351372 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ltqf4\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.372448 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ltqf4\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.374049 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffqr8\" (UniqueName: \"kubernetes.io/projected/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-kube-api-access-ffqr8\") pod \"ssh-known-hosts-edpm-deployment-ltqf4\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:04 crc kubenswrapper[4767]: I0317 16:18:04.516102 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:05 crc kubenswrapper[4767]: I0317 16:18:05.123969 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ltqf4"] Mar 17 16:18:05 crc kubenswrapper[4767]: I0317 16:18:05.579604 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562738-k77dn" Mar 17 16:18:05 crc kubenswrapper[4767]: I0317 16:18:05.670276 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmgl5\" (UniqueName: \"kubernetes.io/projected/8bd93414-e722-485f-a4e5-453ef7d6c4aa-kube-api-access-zmgl5\") pod \"8bd93414-e722-485f-a4e5-453ef7d6c4aa\" (UID: \"8bd93414-e722-485f-a4e5-453ef7d6c4aa\") " Mar 17 16:18:05 crc kubenswrapper[4767]: I0317 16:18:05.678810 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bd93414-e722-485f-a4e5-453ef7d6c4aa-kube-api-access-zmgl5" (OuterVolumeSpecName: "kube-api-access-zmgl5") pod "8bd93414-e722-485f-a4e5-453ef7d6c4aa" (UID: "8bd93414-e722-485f-a4e5-453ef7d6c4aa"). InnerVolumeSpecName "kube-api-access-zmgl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:18:05 crc kubenswrapper[4767]: I0317 16:18:05.772672 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmgl5\" (UniqueName: \"kubernetes.io/projected/8bd93414-e722-485f-a4e5-453ef7d6c4aa-kube-api-access-zmgl5\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:06 crc kubenswrapper[4767]: I0317 16:18:06.142361 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562738-k77dn" event={"ID":"8bd93414-e722-485f-a4e5-453ef7d6c4aa","Type":"ContainerDied","Data":"14716ede3f69c7be204b009b7c62a2e1285450dcd5943068d91152894f638b38"} Mar 17 16:18:06 crc kubenswrapper[4767]: I0317 16:18:06.142821 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14716ede3f69c7be204b009b7c62a2e1285450dcd5943068d91152894f638b38" Mar 17 16:18:06 crc kubenswrapper[4767]: I0317 16:18:06.142454 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562738-k77dn" Mar 17 16:18:06 crc kubenswrapper[4767]: I0317 16:18:06.145377 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" event={"ID":"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b","Type":"ContainerStarted","Data":"507398b3fce7829ad390a2fc4d3c1b2c746dc679bdf9b8c7e32f3e9cc892865c"} Mar 17 16:18:06 crc kubenswrapper[4767]: I0317 16:18:06.310295 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562732-s2frn"] Mar 17 16:18:06 crc kubenswrapper[4767]: I0317 16:18:06.327444 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562732-s2frn"] Mar 17 16:18:07 crc kubenswrapper[4767]: I0317 16:18:07.158313 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" event={"ID":"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b","Type":"ContainerStarted","Data":"e0dd6f81eff9df64ea046d93c154d3a7c3db46f5c28d6677a91abee06e964698"} Mar 17 16:18:07 crc kubenswrapper[4767]: I0317 16:18:07.178466 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" podStartSLOduration=1.8665990190000001 podStartE2EDuration="3.178445476s" podCreationTimestamp="2026-03-17 16:18:04 +0000 UTC" firstStartedPulling="2026-03-17 16:18:05.14381651 +0000 UTC m=+2476.557132557" lastFinishedPulling="2026-03-17 16:18:06.455662967 +0000 UTC m=+2477.868979014" observedRunningTime="2026-03-17 16:18:07.175866141 +0000 UTC m=+2478.589182208" watchObservedRunningTime="2026-03-17 16:18:07.178445476 +0000 UTC m=+2478.591761523" Mar 17 16:18:07 crc kubenswrapper[4767]: I0317 16:18:07.355760 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:18:07 crc kubenswrapper[4767]: E0317 16:18:07.356114 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:18:07 crc kubenswrapper[4767]: I0317 16:18:07.369689 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af290f7-a893-41d9-87ce-4e6f42d73e28" path="/var/lib/kubelet/pods/1af290f7-a893-41d9-87ce-4e6f42d73e28/volumes" Mar 17 16:18:14 crc kubenswrapper[4767]: I0317 16:18:14.367866 4767 generic.go:334] "Generic (PLEG): container finished" podID="d6b74dd7-e81d-48c3-8b5c-1e684d32e64b" containerID="e0dd6f81eff9df64ea046d93c154d3a7c3db46f5c28d6677a91abee06e964698" exitCode=0 Mar 17 16:18:14 crc kubenswrapper[4767]: I0317 16:18:14.367975 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" event={"ID":"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b","Type":"ContainerDied","Data":"e0dd6f81eff9df64ea046d93c154d3a7c3db46f5c28d6677a91abee06e964698"} Mar 17 16:18:15 crc kubenswrapper[4767]: I0317 16:18:15.888613 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.002828 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffqr8\" (UniqueName: \"kubernetes.io/projected/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-kube-api-access-ffqr8\") pod \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.003706 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-ssh-key-openstack-edpm-ipam\") pod \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.004077 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-inventory-0\") pod \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\" (UID: \"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b\") " Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.010536 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-kube-api-access-ffqr8" (OuterVolumeSpecName: "kube-api-access-ffqr8") pod "d6b74dd7-e81d-48c3-8b5c-1e684d32e64b" (UID: "d6b74dd7-e81d-48c3-8b5c-1e684d32e64b"). InnerVolumeSpecName "kube-api-access-ffqr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.037480 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "d6b74dd7-e81d-48c3-8b5c-1e684d32e64b" (UID: "d6b74dd7-e81d-48c3-8b5c-1e684d32e64b"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.055617 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d6b74dd7-e81d-48c3-8b5c-1e684d32e64b" (UID: "d6b74dd7-e81d-48c3-8b5c-1e684d32e64b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.120666 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.120722 4767 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.120739 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffqr8\" (UniqueName: \"kubernetes.io/projected/d6b74dd7-e81d-48c3-8b5c-1e684d32e64b-kube-api-access-ffqr8\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.397101 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" event={"ID":"d6b74dd7-e81d-48c3-8b5c-1e684d32e64b","Type":"ContainerDied","Data":"507398b3fce7829ad390a2fc4d3c1b2c746dc679bdf9b8c7e32f3e9cc892865c"} Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.397153 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="507398b3fce7829ad390a2fc4d3c1b2c746dc679bdf9b8c7e32f3e9cc892865c" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.397198 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ltqf4" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.508778 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w"] Mar 17 16:18:16 crc kubenswrapper[4767]: E0317 16:18:16.510352 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b74dd7-e81d-48c3-8b5c-1e684d32e64b" containerName="ssh-known-hosts-edpm-deployment" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.510384 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b74dd7-e81d-48c3-8b5c-1e684d32e64b" containerName="ssh-known-hosts-edpm-deployment" Mar 17 16:18:16 crc kubenswrapper[4767]: E0317 16:18:16.510450 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd93414-e722-485f-a4e5-453ef7d6c4aa" containerName="oc" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.510459 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd93414-e722-485f-a4e5-453ef7d6c4aa" containerName="oc" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.510773 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6b74dd7-e81d-48c3-8b5c-1e684d32e64b" containerName="ssh-known-hosts-edpm-deployment" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.510815 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bd93414-e722-485f-a4e5-453ef7d6c4aa" containerName="oc" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.515492 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.518866 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.518866 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.518908 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.519533 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.531304 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w"] Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.865638 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-drn8w\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.865701 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkrsj\" (UniqueName: \"kubernetes.io/projected/b9540b6a-23ed-4dd0-9294-901d6ddfe539-kube-api-access-dkrsj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-drn8w\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.866131 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-drn8w\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.969151 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-drn8w\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.969239 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-drn8w\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.969279 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkrsj\" (UniqueName: \"kubernetes.io/projected/b9540b6a-23ed-4dd0-9294-901d6ddfe539-kube-api-access-dkrsj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-drn8w\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.977587 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-drn8w\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:16 crc kubenswrapper[4767]: I0317 16:18:16.994667 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-drn8w\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:17 crc kubenswrapper[4767]: I0317 16:18:17.001825 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkrsj\" (UniqueName: \"kubernetes.io/projected/b9540b6a-23ed-4dd0-9294-901d6ddfe539-kube-api-access-dkrsj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-drn8w\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:17 crc kubenswrapper[4767]: I0317 16:18:17.151250 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:17 crc kubenswrapper[4767]: I0317 16:18:17.776656 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w"] Mar 17 16:18:18 crc kubenswrapper[4767]: I0317 16:18:18.439426 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" event={"ID":"b9540b6a-23ed-4dd0-9294-901d6ddfe539","Type":"ContainerStarted","Data":"c642defcd30753ce77f438bdeb835fcb10b83afde3d1c4d75f77877399872dd5"} Mar 17 16:18:19 crc kubenswrapper[4767]: I0317 16:18:19.452848 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" event={"ID":"b9540b6a-23ed-4dd0-9294-901d6ddfe539","Type":"ContainerStarted","Data":"7e6a802ab585321a890ec38ffaa43f0adf32274bacbd55e723bd6085a6755d9e"} Mar 17 16:18:19 crc kubenswrapper[4767]: I0317 16:18:19.508511 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" podStartSLOduration=2.966136172 podStartE2EDuration="3.508477728s" podCreationTimestamp="2026-03-17 16:18:16 +0000 UTC" firstStartedPulling="2026-03-17 16:18:17.782203056 +0000 UTC m=+2489.195519093" lastFinishedPulling="2026-03-17 16:18:18.324544602 +0000 UTC m=+2489.737860649" observedRunningTime="2026-03-17 16:18:19.49542757 +0000 UTC m=+2490.908743637" watchObservedRunningTime="2026-03-17 16:18:19.508477728 +0000 UTC m=+2490.921793775" Mar 17 16:18:20 crc kubenswrapper[4767]: I0317 16:18:20.511545 4767 scope.go:117] "RemoveContainer" containerID="e6cf33aa20c58d913e829a4d89bff396eb3e049dc1a45765a5d217387eebeb29" Mar 17 16:18:22 crc kubenswrapper[4767]: I0317 16:18:22.355373 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:18:22 crc kubenswrapper[4767]: E0317 16:18:22.357274 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:18:26 crc kubenswrapper[4767]: I0317 16:18:26.583432 4767 generic.go:334] "Generic (PLEG): container finished" podID="b9540b6a-23ed-4dd0-9294-901d6ddfe539" containerID="7e6a802ab585321a890ec38ffaa43f0adf32274bacbd55e723bd6085a6755d9e" exitCode=0 Mar 17 16:18:26 crc kubenswrapper[4767]: I0317 16:18:26.583504 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" event={"ID":"b9540b6a-23ed-4dd0-9294-901d6ddfe539","Type":"ContainerDied","Data":"7e6a802ab585321a890ec38ffaa43f0adf32274bacbd55e723bd6085a6755d9e"} Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.136853 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.310765 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-ssh-key-openstack-edpm-ipam\") pod \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.310863 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkrsj\" (UniqueName: \"kubernetes.io/projected/b9540b6a-23ed-4dd0-9294-901d6ddfe539-kube-api-access-dkrsj\") pod \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.311003 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-inventory\") pod \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\" (UID: \"b9540b6a-23ed-4dd0-9294-901d6ddfe539\") " Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.316930 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9540b6a-23ed-4dd0-9294-901d6ddfe539-kube-api-access-dkrsj" (OuterVolumeSpecName: "kube-api-access-dkrsj") pod "b9540b6a-23ed-4dd0-9294-901d6ddfe539" (UID: "b9540b6a-23ed-4dd0-9294-901d6ddfe539"). InnerVolumeSpecName "kube-api-access-dkrsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.345385 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-inventory" (OuterVolumeSpecName: "inventory") pod "b9540b6a-23ed-4dd0-9294-901d6ddfe539" (UID: "b9540b6a-23ed-4dd0-9294-901d6ddfe539"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.347135 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b9540b6a-23ed-4dd0-9294-901d6ddfe539" (UID: "b9540b6a-23ed-4dd0-9294-901d6ddfe539"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.414853 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkrsj\" (UniqueName: \"kubernetes.io/projected/b9540b6a-23ed-4dd0-9294-901d6ddfe539-kube-api-access-dkrsj\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.414909 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.414923 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b9540b6a-23ed-4dd0-9294-901d6ddfe539-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.625472 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" event={"ID":"b9540b6a-23ed-4dd0-9294-901d6ddfe539","Type":"ContainerDied","Data":"c642defcd30753ce77f438bdeb835fcb10b83afde3d1c4d75f77877399872dd5"} Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.625539 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c642defcd30753ce77f438bdeb835fcb10b83afde3d1c4d75f77877399872dd5" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.626227 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-drn8w" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.711223 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5"] Mar 17 16:18:28 crc kubenswrapper[4767]: E0317 16:18:28.712519 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9540b6a-23ed-4dd0-9294-901d6ddfe539" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.712566 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9540b6a-23ed-4dd0-9294-901d6ddfe539" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.712983 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9540b6a-23ed-4dd0-9294-901d6ddfe539" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.714824 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.718228 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.719224 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.719476 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.721674 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.732357 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5"] Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.756383 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.756865 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.757292 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgmqs\" (UniqueName: \"kubernetes.io/projected/9963968d-ff4e-4179-9b26-504bcf4748e9-kube-api-access-sgmqs\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.859700 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgmqs\" (UniqueName: \"kubernetes.io/projected/9963968d-ff4e-4179-9b26-504bcf4748e9-kube-api-access-sgmqs\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.859992 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.860165 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.864231 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.864592 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:28 crc kubenswrapper[4767]: I0317 16:18:28.879191 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgmqs\" (UniqueName: \"kubernetes.io/projected/9963968d-ff4e-4179-9b26-504bcf4748e9-kube-api-access-sgmqs\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:29 crc kubenswrapper[4767]: I0317 16:18:29.060550 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:29 crc kubenswrapper[4767]: I0317 16:18:29.701605 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5"] Mar 17 16:18:30 crc kubenswrapper[4767]: I0317 16:18:30.652761 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" event={"ID":"9963968d-ff4e-4179-9b26-504bcf4748e9","Type":"ContainerStarted","Data":"a3ac0d29e777eb0962119dbd1c751d4a666fd24e8bb48a54666e625ca36f4d48"} Mar 17 16:18:30 crc kubenswrapper[4767]: I0317 16:18:30.653123 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" event={"ID":"9963968d-ff4e-4179-9b26-504bcf4748e9","Type":"ContainerStarted","Data":"50c432bab09b7ca1880595d3558212c8793c24de94fdd4d201fcd3ad8805ea9d"} Mar 17 16:18:30 crc kubenswrapper[4767]: I0317 16:18:30.680665 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" podStartSLOduration=2.280722824 podStartE2EDuration="2.680636079s" podCreationTimestamp="2026-03-17 16:18:28 +0000 UTC" firstStartedPulling="2026-03-17 16:18:29.718938485 +0000 UTC m=+2501.132254532" lastFinishedPulling="2026-03-17 16:18:30.11885173 +0000 UTC m=+2501.532167787" observedRunningTime="2026-03-17 16:18:30.671512295 +0000 UTC m=+2502.084828342" watchObservedRunningTime="2026-03-17 16:18:30.680636079 +0000 UTC m=+2502.093952126" Mar 17 16:18:36 crc kubenswrapper[4767]: I0317 16:18:36.355289 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:18:36 crc kubenswrapper[4767]: E0317 16:18:36.356216 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:18:39 crc kubenswrapper[4767]: I0317 16:18:39.768931 4767 generic.go:334] "Generic (PLEG): container finished" podID="9963968d-ff4e-4179-9b26-504bcf4748e9" containerID="a3ac0d29e777eb0962119dbd1c751d4a666fd24e8bb48a54666e625ca36f4d48" exitCode=0 Mar 17 16:18:39 crc kubenswrapper[4767]: I0317 16:18:39.769018 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" event={"ID":"9963968d-ff4e-4179-9b26-504bcf4748e9","Type":"ContainerDied","Data":"a3ac0d29e777eb0962119dbd1c751d4a666fd24e8bb48a54666e625ca36f4d48"} Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.297804 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.486404 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-inventory\") pod \"9963968d-ff4e-4179-9b26-504bcf4748e9\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.487036 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-ssh-key-openstack-edpm-ipam\") pod \"9963968d-ff4e-4179-9b26-504bcf4748e9\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.487073 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgmqs\" (UniqueName: \"kubernetes.io/projected/9963968d-ff4e-4179-9b26-504bcf4748e9-kube-api-access-sgmqs\") pod \"9963968d-ff4e-4179-9b26-504bcf4748e9\" (UID: \"9963968d-ff4e-4179-9b26-504bcf4748e9\") " Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.492505 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9963968d-ff4e-4179-9b26-504bcf4748e9-kube-api-access-sgmqs" (OuterVolumeSpecName: "kube-api-access-sgmqs") pod "9963968d-ff4e-4179-9b26-504bcf4748e9" (UID: "9963968d-ff4e-4179-9b26-504bcf4748e9"). InnerVolumeSpecName "kube-api-access-sgmqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.526216 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9963968d-ff4e-4179-9b26-504bcf4748e9" (UID: "9963968d-ff4e-4179-9b26-504bcf4748e9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.529144 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-inventory" (OuterVolumeSpecName: "inventory") pod "9963968d-ff4e-4179-9b26-504bcf4748e9" (UID: "9963968d-ff4e-4179-9b26-504bcf4748e9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.591042 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.591089 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgmqs\" (UniqueName: \"kubernetes.io/projected/9963968d-ff4e-4179-9b26-504bcf4748e9-kube-api-access-sgmqs\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.591145 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9963968d-ff4e-4179-9b26-504bcf4748e9-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.795722 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" event={"ID":"9963968d-ff4e-4179-9b26-504bcf4748e9","Type":"ContainerDied","Data":"50c432bab09b7ca1880595d3558212c8793c24de94fdd4d201fcd3ad8805ea9d"} Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.795764 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50c432bab09b7ca1880595d3558212c8793c24de94fdd4d201fcd3ad8805ea9d" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.795800 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dtvw5" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.903552 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr"] Mar 17 16:18:41 crc kubenswrapper[4767]: E0317 16:18:41.904693 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9963968d-ff4e-4179-9b26-504bcf4748e9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.904889 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="9963968d-ff4e-4179-9b26-504bcf4748e9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.907209 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="9963968d-ff4e-4179-9b26-504bcf4748e9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.908491 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.912071 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.913010 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.913264 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.913354 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.914409 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.914560 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.914707 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.914739 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.915050 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:18:41 crc kubenswrapper[4767]: I0317 16:18:41.917338 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr"] Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.104654 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.104722 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.104851 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9bzr\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-kube-api-access-m9bzr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.104916 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.104976 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.105013 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.105128 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.105250 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.105298 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.105378 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.105476 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.105546 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.105586 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.105627 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.105676 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.105716 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.208472 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.208835 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.208889 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.208923 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.209517 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.209580 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.209651 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.209689 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.209740 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.209779 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.209810 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.209865 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.209906 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.209970 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9bzr\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-kube-api-access-m9bzr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.210084 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.210142 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.214619 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.215125 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.215196 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.215318 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.215518 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.216936 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.218090 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.219512 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.219689 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.219687 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.219820 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.219970 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.220494 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.220518 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.222574 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.230476 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9bzr\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-kube-api-access-m9bzr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:42 crc kubenswrapper[4767]: I0317 16:18:42.530745 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:18:43 crc kubenswrapper[4767]: I0317 16:18:43.116311 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr"] Mar 17 16:18:43 crc kubenswrapper[4767]: W0317 16:18:43.120133 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5e5e38f_ce87_43ae_83ef_73d341d04c07.slice/crio-79bfe83c61004efed4c37bc887ae7159833935dd50fa41bdfc5c28ac77bb7963 WatchSource:0}: Error finding container 79bfe83c61004efed4c37bc887ae7159833935dd50fa41bdfc5c28ac77bb7963: Status 404 returned error can't find the container with id 79bfe83c61004efed4c37bc887ae7159833935dd50fa41bdfc5c28ac77bb7963 Mar 17 16:18:43 crc kubenswrapper[4767]: I0317 16:18:43.863412 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" event={"ID":"f5e5e38f-ce87-43ae-83ef-73d341d04c07","Type":"ContainerStarted","Data":"79bfe83c61004efed4c37bc887ae7159833935dd50fa41bdfc5c28ac77bb7963"} Mar 17 16:18:44 crc kubenswrapper[4767]: I0317 16:18:44.875727 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" event={"ID":"f5e5e38f-ce87-43ae-83ef-73d341d04c07","Type":"ContainerStarted","Data":"1f205d97c24e15de1141ec6f9a2ef0237f109399e31d48ad85adc209595c45da"} Mar 17 16:18:44 crc kubenswrapper[4767]: I0317 16:18:44.898129 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" podStartSLOduration=2.639648423 podStartE2EDuration="3.898110535s" podCreationTimestamp="2026-03-17 16:18:41 +0000 UTC" firstStartedPulling="2026-03-17 16:18:43.125403019 +0000 UTC m=+2514.538719056" lastFinishedPulling="2026-03-17 16:18:44.383865131 +0000 UTC m=+2515.797181168" observedRunningTime="2026-03-17 16:18:44.896073946 +0000 UTC m=+2516.309390003" watchObservedRunningTime="2026-03-17 16:18:44.898110535 +0000 UTC m=+2516.311426582" Mar 17 16:18:51 crc kubenswrapper[4767]: I0317 16:18:51.355430 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:18:51 crc kubenswrapper[4767]: E0317 16:18:51.356470 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:18:57 crc kubenswrapper[4767]: I0317 16:18:57.050992 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-4zghh"] Mar 17 16:18:57 crc kubenswrapper[4767]: I0317 16:18:57.068703 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-4zghh"] Mar 17 16:18:57 crc kubenswrapper[4767]: I0317 16:18:57.371926 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d28d7ecb-5431-4abd-beb2-f4a4c22df3f3" path="/var/lib/kubelet/pods/d28d7ecb-5431-4abd-beb2-f4a4c22df3f3/volumes" Mar 17 16:19:03 crc kubenswrapper[4767]: I0317 16:19:03.354853 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:19:03 crc kubenswrapper[4767]: E0317 16:19:03.355687 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:19:18 crc kubenswrapper[4767]: I0317 16:19:18.355257 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:19:18 crc kubenswrapper[4767]: E0317 16:19:18.356132 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:19:20 crc kubenswrapper[4767]: I0317 16:19:20.618163 4767 scope.go:117] "RemoveContainer" containerID="da6c402c498e398f77675e3566c0f0e0d201675f910e2c5130652af46c55ad1f" Mar 17 16:19:25 crc kubenswrapper[4767]: I0317 16:19:25.374200 4767 generic.go:334] "Generic (PLEG): container finished" podID="f5e5e38f-ce87-43ae-83ef-73d341d04c07" containerID="1f205d97c24e15de1141ec6f9a2ef0237f109399e31d48ad85adc209595c45da" exitCode=0 Mar 17 16:19:25 crc kubenswrapper[4767]: I0317 16:19:25.374495 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" event={"ID":"f5e5e38f-ce87-43ae-83ef-73d341d04c07","Type":"ContainerDied","Data":"1f205d97c24e15de1141ec6f9a2ef0237f109399e31d48ad85adc209595c45da"} Mar 17 16:19:26 crc kubenswrapper[4767]: I0317 16:19:26.967104 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125390 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125460 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-repo-setup-combined-ca-bundle\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125518 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-bootstrap-combined-ca-bundle\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125536 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-combined-ca-bundle\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125566 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-neutron-metadata-combined-ca-bundle\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125589 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125734 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-power-monitoring-combined-ca-bundle\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125779 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125810 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125834 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ovn-combined-ca-bundle\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125853 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ssh-key-openstack-edpm-ipam\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125898 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-ovn-default-certs-0\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125926 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-inventory\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.125952 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9bzr\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-kube-api-access-m9bzr\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.126088 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-nova-combined-ca-bundle\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.126123 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-libvirt-combined-ca-bundle\") pod \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\" (UID: \"f5e5e38f-ce87-43ae-83ef-73d341d04c07\") " Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.133517 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.133986 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.134508 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.136011 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.136247 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.136868 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.136925 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.137407 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.138123 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.138542 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-kube-api-access-m9bzr" (OuterVolumeSpecName: "kube-api-access-m9bzr") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "kube-api-access-m9bzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.139225 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.139891 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.139965 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.144525 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.169449 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-inventory" (OuterVolumeSpecName: "inventory") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.169858 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f5e5e38f-ce87-43ae-83ef-73d341d04c07" (UID: "f5e5e38f-ce87-43ae-83ef-73d341d04c07"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.230730 4767 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231002 4767 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231014 4767 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231023 4767 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231032 4767 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231041 4767 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231053 4767 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231071 4767 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231082 4767 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231094 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231104 4767 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231112 4767 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231123 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231132 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9bzr\" (UniqueName: \"kubernetes.io/projected/f5e5e38f-ce87-43ae-83ef-73d341d04c07-kube-api-access-m9bzr\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231141 4767 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.231149 4767 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e5e38f-ce87-43ae-83ef-73d341d04c07-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.415221 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" event={"ID":"f5e5e38f-ce87-43ae-83ef-73d341d04c07","Type":"ContainerDied","Data":"79bfe83c61004efed4c37bc887ae7159833935dd50fa41bdfc5c28ac77bb7963"} Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.415271 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79bfe83c61004efed4c37bc887ae7159833935dd50fa41bdfc5c28ac77bb7963" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.415379 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4lzlr" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.743020 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw"] Mar 17 16:19:27 crc kubenswrapper[4767]: E0317 16:19:27.744243 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e5e38f-ce87-43ae-83ef-73d341d04c07" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.744270 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e5e38f-ce87-43ae-83ef-73d341d04c07" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.744642 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5e5e38f-ce87-43ae-83ef-73d341d04c07" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.746124 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.749190 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.749437 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.749614 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.750476 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.757911 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw"] Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.773981 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.878024 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.878159 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwrwl\" (UniqueName: \"kubernetes.io/projected/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-kube-api-access-cwrwl\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.878240 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.878311 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.879935 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.983250 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.983804 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwrwl\" (UniqueName: \"kubernetes.io/projected/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-kube-api-access-cwrwl\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.983864 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.983970 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.984028 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.985142 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.992352 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.993485 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:27 crc kubenswrapper[4767]: I0317 16:19:27.998198 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:28 crc kubenswrapper[4767]: I0317 16:19:28.010015 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwrwl\" (UniqueName: \"kubernetes.io/projected/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-kube-api-access-cwrwl\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-zqbsw\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:28 crc kubenswrapper[4767]: I0317 16:19:28.100235 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:19:28 crc kubenswrapper[4767]: I0317 16:19:28.703450 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw"] Mar 17 16:19:29 crc kubenswrapper[4767]: I0317 16:19:29.438628 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" event={"ID":"46cb2975-a7e0-4fe7-8f91-8194638b0c1f","Type":"ContainerStarted","Data":"903b6991a72997a5a0d659da90a7ce27af0c6725d07ea7083a041a68ae8ca1ed"} Mar 17 16:19:30 crc kubenswrapper[4767]: I0317 16:19:30.452981 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" event={"ID":"46cb2975-a7e0-4fe7-8f91-8194638b0c1f","Type":"ContainerStarted","Data":"53b23758c3b373d4105b45a30866c726aa197107ef06fe85f3a47805fc2a4d7f"} Mar 17 16:19:30 crc kubenswrapper[4767]: I0317 16:19:30.485401 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" podStartSLOduration=2.8452197359999998 podStartE2EDuration="3.485369113s" podCreationTimestamp="2026-03-17 16:19:27 +0000 UTC" firstStartedPulling="2026-03-17 16:19:28.698156758 +0000 UTC m=+2560.111472805" lastFinishedPulling="2026-03-17 16:19:29.338306135 +0000 UTC m=+2560.751622182" observedRunningTime="2026-03-17 16:19:30.476736443 +0000 UTC m=+2561.890052490" watchObservedRunningTime="2026-03-17 16:19:30.485369113 +0000 UTC m=+2561.898685160" Mar 17 16:19:32 crc kubenswrapper[4767]: I0317 16:19:32.354675 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:19:32 crc kubenswrapper[4767]: E0317 16:19:32.355106 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:19:40 crc kubenswrapper[4767]: I0317 16:19:40.946312 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hrvvh"] Mar 17 16:19:40 crc kubenswrapper[4767]: I0317 16:19:40.950139 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:40 crc kubenswrapper[4767]: I0317 16:19:40.963444 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrvvh"] Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.050512 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-c9h7f"] Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.066953 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-c9h7f"] Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.118191 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89h8x\" (UniqueName: \"kubernetes.io/projected/5af033a0-6d5d-4719-b732-aa72dcd4c883-kube-api-access-89h8x\") pod \"redhat-operators-hrvvh\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.118578 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-utilities\") pod \"redhat-operators-hrvvh\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.118849 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-catalog-content\") pod \"redhat-operators-hrvvh\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.222240 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-utilities\") pod \"redhat-operators-hrvvh\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.222359 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-catalog-content\") pod \"redhat-operators-hrvvh\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.222569 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89h8x\" (UniqueName: \"kubernetes.io/projected/5af033a0-6d5d-4719-b732-aa72dcd4c883-kube-api-access-89h8x\") pod \"redhat-operators-hrvvh\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.222865 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-utilities\") pod \"redhat-operators-hrvvh\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.223114 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-catalog-content\") pod \"redhat-operators-hrvvh\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.251437 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89h8x\" (UniqueName: \"kubernetes.io/projected/5af033a0-6d5d-4719-b732-aa72dcd4c883-kube-api-access-89h8x\") pod \"redhat-operators-hrvvh\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.278675 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.373738 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2" path="/var/lib/kubelet/pods/fdaa047c-3098-4ee4-b6bf-ffadc6d6fcb2/volumes" Mar 17 16:19:41 crc kubenswrapper[4767]: I0317 16:19:41.843294 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrvvh"] Mar 17 16:19:42 crc kubenswrapper[4767]: I0317 16:19:42.802054 4767 generic.go:334] "Generic (PLEG): container finished" podID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerID="9a0d6efbee1e2fb191c668fc40e756246095914a3f26ba44359c153a6c1e605e" exitCode=0 Mar 17 16:19:42 crc kubenswrapper[4767]: I0317 16:19:42.802118 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvvh" event={"ID":"5af033a0-6d5d-4719-b732-aa72dcd4c883","Type":"ContainerDied","Data":"9a0d6efbee1e2fb191c668fc40e756246095914a3f26ba44359c153a6c1e605e"} Mar 17 16:19:42 crc kubenswrapper[4767]: I0317 16:19:42.802157 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvvh" event={"ID":"5af033a0-6d5d-4719-b732-aa72dcd4c883","Type":"ContainerStarted","Data":"8ec0cd2f532834fe65aee49e4c7cca5cfb776cf4da11b808be1b12d94f248e04"} Mar 17 16:19:44 crc kubenswrapper[4767]: I0317 16:19:44.828090 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvvh" event={"ID":"5af033a0-6d5d-4719-b732-aa72dcd4c883","Type":"ContainerStarted","Data":"8989f608bfead243dd656b5497f9ca48c02f32af1feb5addda924e392be96cbd"} Mar 17 16:19:46 crc kubenswrapper[4767]: I0317 16:19:46.354969 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:19:46 crc kubenswrapper[4767]: E0317 16:19:46.357465 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:19:49 crc kubenswrapper[4767]: I0317 16:19:49.194710 4767 generic.go:334] "Generic (PLEG): container finished" podID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerID="8989f608bfead243dd656b5497f9ca48c02f32af1feb5addda924e392be96cbd" exitCode=0 Mar 17 16:19:49 crc kubenswrapper[4767]: I0317 16:19:49.194806 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvvh" event={"ID":"5af033a0-6d5d-4719-b732-aa72dcd4c883","Type":"ContainerDied","Data":"8989f608bfead243dd656b5497f9ca48c02f32af1feb5addda924e392be96cbd"} Mar 17 16:19:50 crc kubenswrapper[4767]: I0317 16:19:50.212522 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvvh" event={"ID":"5af033a0-6d5d-4719-b732-aa72dcd4c883","Type":"ContainerStarted","Data":"d144e6304178c5070b415dccbf14157cd530e920a985aba57951dd4be2de52bd"} Mar 17 16:19:50 crc kubenswrapper[4767]: I0317 16:19:50.254864 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hrvvh" podStartSLOduration=3.426452658 podStartE2EDuration="10.254833992s" podCreationTimestamp="2026-03-17 16:19:40 +0000 UTC" firstStartedPulling="2026-03-17 16:19:42.804726905 +0000 UTC m=+2574.218042952" lastFinishedPulling="2026-03-17 16:19:49.633108239 +0000 UTC m=+2581.046424286" observedRunningTime="2026-03-17 16:19:50.236681607 +0000 UTC m=+2581.649997654" watchObservedRunningTime="2026-03-17 16:19:50.254833992 +0000 UTC m=+2581.668150049" Mar 17 16:19:51 crc kubenswrapper[4767]: I0317 16:19:51.279429 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:51 crc kubenswrapper[4767]: I0317 16:19:51.280253 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:19:52 crc kubenswrapper[4767]: I0317 16:19:52.329901 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hrvvh" podUID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerName="registry-server" probeResult="failure" output=< Mar 17 16:19:52 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:19:52 crc kubenswrapper[4767]: > Mar 17 16:20:00 crc kubenswrapper[4767]: I0317 16:20:00.160876 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562740-cv9nj"] Mar 17 16:20:00 crc kubenswrapper[4767]: I0317 16:20:00.164118 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562740-cv9nj" Mar 17 16:20:00 crc kubenswrapper[4767]: I0317 16:20:00.169159 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:20:00 crc kubenswrapper[4767]: I0317 16:20:00.172992 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562740-cv9nj"] Mar 17 16:20:00 crc kubenswrapper[4767]: I0317 16:20:00.175561 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:20:00 crc kubenswrapper[4767]: I0317 16:20:00.178594 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:20:00 crc kubenswrapper[4767]: I0317 16:20:00.195234 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4nkc\" (UniqueName: \"kubernetes.io/projected/8411da78-bfed-47a5-8e0d-1ed245fecd48-kube-api-access-v4nkc\") pod \"auto-csr-approver-29562740-cv9nj\" (UID: \"8411da78-bfed-47a5-8e0d-1ed245fecd48\") " pod="openshift-infra/auto-csr-approver-29562740-cv9nj" Mar 17 16:20:00 crc kubenswrapper[4767]: I0317 16:20:00.296681 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4nkc\" (UniqueName: \"kubernetes.io/projected/8411da78-bfed-47a5-8e0d-1ed245fecd48-kube-api-access-v4nkc\") pod \"auto-csr-approver-29562740-cv9nj\" (UID: \"8411da78-bfed-47a5-8e0d-1ed245fecd48\") " pod="openshift-infra/auto-csr-approver-29562740-cv9nj" Mar 17 16:20:00 crc kubenswrapper[4767]: I0317 16:20:00.337960 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4nkc\" (UniqueName: \"kubernetes.io/projected/8411da78-bfed-47a5-8e0d-1ed245fecd48-kube-api-access-v4nkc\") pod \"auto-csr-approver-29562740-cv9nj\" (UID: \"8411da78-bfed-47a5-8e0d-1ed245fecd48\") " pod="openshift-infra/auto-csr-approver-29562740-cv9nj" Mar 17 16:20:00 crc kubenswrapper[4767]: I0317 16:20:00.490033 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562740-cv9nj" Mar 17 16:20:01 crc kubenswrapper[4767]: I0317 16:20:01.185257 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562740-cv9nj"] Mar 17 16:20:01 crc kubenswrapper[4767]: I0317 16:20:01.355762 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:20:01 crc kubenswrapper[4767]: E0317 16:20:01.356247 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:20:01 crc kubenswrapper[4767]: I0317 16:20:01.369341 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562740-cv9nj" event={"ID":"8411da78-bfed-47a5-8e0d-1ed245fecd48","Type":"ContainerStarted","Data":"a7a5d984bbcba5fc723d9a480a20212230d859f67d817461e3a2a0050405ffe5"} Mar 17 16:20:02 crc kubenswrapper[4767]: I0317 16:20:02.338410 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hrvvh" podUID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerName="registry-server" probeResult="failure" output=< Mar 17 16:20:02 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:20:02 crc kubenswrapper[4767]: > Mar 17 16:20:04 crc kubenswrapper[4767]: I0317 16:20:04.522094 4767 generic.go:334] "Generic (PLEG): container finished" podID="8411da78-bfed-47a5-8e0d-1ed245fecd48" containerID="2ace0c423c7d66ab119f54c12faa314382cdc010e67e852f7367de8d77c9cab1" exitCode=0 Mar 17 16:20:04 crc kubenswrapper[4767]: I0317 16:20:04.522200 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562740-cv9nj" event={"ID":"8411da78-bfed-47a5-8e0d-1ed245fecd48","Type":"ContainerDied","Data":"2ace0c423c7d66ab119f54c12faa314382cdc010e67e852f7367de8d77c9cab1"} Mar 17 16:20:05 crc kubenswrapper[4767]: I0317 16:20:05.969396 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562740-cv9nj" Mar 17 16:20:06 crc kubenswrapper[4767]: I0317 16:20:06.088810 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4nkc\" (UniqueName: \"kubernetes.io/projected/8411da78-bfed-47a5-8e0d-1ed245fecd48-kube-api-access-v4nkc\") pod \"8411da78-bfed-47a5-8e0d-1ed245fecd48\" (UID: \"8411da78-bfed-47a5-8e0d-1ed245fecd48\") " Mar 17 16:20:06 crc kubenswrapper[4767]: I0317 16:20:06.097939 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8411da78-bfed-47a5-8e0d-1ed245fecd48-kube-api-access-v4nkc" (OuterVolumeSpecName: "kube-api-access-v4nkc") pod "8411da78-bfed-47a5-8e0d-1ed245fecd48" (UID: "8411da78-bfed-47a5-8e0d-1ed245fecd48"). InnerVolumeSpecName "kube-api-access-v4nkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:20:06 crc kubenswrapper[4767]: I0317 16:20:06.195257 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4nkc\" (UniqueName: \"kubernetes.io/projected/8411da78-bfed-47a5-8e0d-1ed245fecd48-kube-api-access-v4nkc\") on node \"crc\" DevicePath \"\"" Mar 17 16:20:06 crc kubenswrapper[4767]: I0317 16:20:06.548918 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562740-cv9nj" event={"ID":"8411da78-bfed-47a5-8e0d-1ed245fecd48","Type":"ContainerDied","Data":"a7a5d984bbcba5fc723d9a480a20212230d859f67d817461e3a2a0050405ffe5"} Mar 17 16:20:06 crc kubenswrapper[4767]: I0317 16:20:06.548971 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7a5d984bbcba5fc723d9a480a20212230d859f67d817461e3a2a0050405ffe5" Mar 17 16:20:06 crc kubenswrapper[4767]: I0317 16:20:06.549735 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562740-cv9nj" Mar 17 16:20:07 crc kubenswrapper[4767]: I0317 16:20:07.053769 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562734-chbkz"] Mar 17 16:20:07 crc kubenswrapper[4767]: I0317 16:20:07.065495 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562734-chbkz"] Mar 17 16:20:07 crc kubenswrapper[4767]: I0317 16:20:07.372827 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18fa314c-78e8-4e3e-a392-855c08a28f62" path="/var/lib/kubelet/pods/18fa314c-78e8-4e3e-a392-855c08a28f62/volumes" Mar 17 16:20:11 crc kubenswrapper[4767]: I0317 16:20:11.334019 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:20:11 crc kubenswrapper[4767]: I0317 16:20:11.399094 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:20:13 crc kubenswrapper[4767]: I0317 16:20:13.472369 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:20:13 crc kubenswrapper[4767]: E0317 16:20:13.473393 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:20:14 crc kubenswrapper[4767]: I0317 16:20:14.252863 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrvvh"] Mar 17 16:20:14 crc kubenswrapper[4767]: I0317 16:20:14.253443 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hrvvh" podUID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerName="registry-server" containerID="cri-o://d144e6304178c5070b415dccbf14157cd530e920a985aba57951dd4be2de52bd" gracePeriod=2 Mar 17 16:20:14 crc kubenswrapper[4767]: I0317 16:20:14.659078 4767 generic.go:334] "Generic (PLEG): container finished" podID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerID="d144e6304178c5070b415dccbf14157cd530e920a985aba57951dd4be2de52bd" exitCode=0 Mar 17 16:20:14 crc kubenswrapper[4767]: I0317 16:20:14.659418 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvvh" event={"ID":"5af033a0-6d5d-4719-b732-aa72dcd4c883","Type":"ContainerDied","Data":"d144e6304178c5070b415dccbf14157cd530e920a985aba57951dd4be2de52bd"} Mar 17 16:20:14 crc kubenswrapper[4767]: I0317 16:20:14.827282 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:20:14 crc kubenswrapper[4767]: I0317 16:20:14.922436 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-utilities\") pod \"5af033a0-6d5d-4719-b732-aa72dcd4c883\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " Mar 17 16:20:14 crc kubenswrapper[4767]: I0317 16:20:14.922550 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89h8x\" (UniqueName: \"kubernetes.io/projected/5af033a0-6d5d-4719-b732-aa72dcd4c883-kube-api-access-89h8x\") pod \"5af033a0-6d5d-4719-b732-aa72dcd4c883\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " Mar 17 16:20:14 crc kubenswrapper[4767]: I0317 16:20:14.923464 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-utilities" (OuterVolumeSpecName: "utilities") pod "5af033a0-6d5d-4719-b732-aa72dcd4c883" (UID: "5af033a0-6d5d-4719-b732-aa72dcd4c883"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:20:14 crc kubenswrapper[4767]: I0317 16:20:14.923952 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-catalog-content\") pod \"5af033a0-6d5d-4719-b732-aa72dcd4c883\" (UID: \"5af033a0-6d5d-4719-b732-aa72dcd4c883\") " Mar 17 16:20:14 crc kubenswrapper[4767]: I0317 16:20:14.925879 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:20:14 crc kubenswrapper[4767]: I0317 16:20:14.931061 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af033a0-6d5d-4719-b732-aa72dcd4c883-kube-api-access-89h8x" (OuterVolumeSpecName: "kube-api-access-89h8x") pod "5af033a0-6d5d-4719-b732-aa72dcd4c883" (UID: "5af033a0-6d5d-4719-b732-aa72dcd4c883"). InnerVolumeSpecName "kube-api-access-89h8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:20:15 crc kubenswrapper[4767]: I0317 16:20:15.028428 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89h8x\" (UniqueName: \"kubernetes.io/projected/5af033a0-6d5d-4719-b732-aa72dcd4c883-kube-api-access-89h8x\") on node \"crc\" DevicePath \"\"" Mar 17 16:20:15 crc kubenswrapper[4767]: I0317 16:20:15.068150 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5af033a0-6d5d-4719-b732-aa72dcd4c883" (UID: "5af033a0-6d5d-4719-b732-aa72dcd4c883"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:20:15 crc kubenswrapper[4767]: I0317 16:20:15.131440 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af033a0-6d5d-4719-b732-aa72dcd4c883-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:20:15 crc kubenswrapper[4767]: I0317 16:20:15.675462 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrvvh" event={"ID":"5af033a0-6d5d-4719-b732-aa72dcd4c883","Type":"ContainerDied","Data":"8ec0cd2f532834fe65aee49e4c7cca5cfb776cf4da11b808be1b12d94f248e04"} Mar 17 16:20:15 crc kubenswrapper[4767]: I0317 16:20:15.675872 4767 scope.go:117] "RemoveContainer" containerID="d144e6304178c5070b415dccbf14157cd530e920a985aba57951dd4be2de52bd" Mar 17 16:20:15 crc kubenswrapper[4767]: I0317 16:20:15.675568 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrvvh" Mar 17 16:20:15 crc kubenswrapper[4767]: I0317 16:20:15.721396 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrvvh"] Mar 17 16:20:15 crc kubenswrapper[4767]: I0317 16:20:15.722515 4767 scope.go:117] "RemoveContainer" containerID="8989f608bfead243dd656b5497f9ca48c02f32af1feb5addda924e392be96cbd" Mar 17 16:20:15 crc kubenswrapper[4767]: I0317 16:20:15.740643 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hrvvh"] Mar 17 16:20:15 crc kubenswrapper[4767]: I0317 16:20:15.757064 4767 scope.go:117] "RemoveContainer" containerID="9a0d6efbee1e2fb191c668fc40e756246095914a3f26ba44359c153a6c1e605e" Mar 17 16:20:17 crc kubenswrapper[4767]: I0317 16:20:17.390223 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5af033a0-6d5d-4719-b732-aa72dcd4c883" path="/var/lib/kubelet/pods/5af033a0-6d5d-4719-b732-aa72dcd4c883/volumes" Mar 17 16:20:20 crc kubenswrapper[4767]: I0317 16:20:20.711482 4767 scope.go:117] "RemoveContainer" containerID="62f8005b8159d57801c533da8eebe0dbed1db72194cbe800b73591ac42cb22b9" Mar 17 16:20:20 crc kubenswrapper[4767]: I0317 16:20:20.779977 4767 scope.go:117] "RemoveContainer" containerID="83f91c392163db2a58f38aab200eec75baf8e6ac41502ce4b37edadf01556977" Mar 17 16:20:28 crc kubenswrapper[4767]: I0317 16:20:28.357122 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:20:28 crc kubenswrapper[4767]: E0317 16:20:28.358006 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:20:32 crc kubenswrapper[4767]: I0317 16:20:32.901099 4767 generic.go:334] "Generic (PLEG): container finished" podID="46cb2975-a7e0-4fe7-8f91-8194638b0c1f" containerID="53b23758c3b373d4105b45a30866c726aa197107ef06fe85f3a47805fc2a4d7f" exitCode=0 Mar 17 16:20:32 crc kubenswrapper[4767]: I0317 16:20:32.901212 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" event={"ID":"46cb2975-a7e0-4fe7-8f91-8194638b0c1f","Type":"ContainerDied","Data":"53b23758c3b373d4105b45a30866c726aa197107ef06fe85f3a47805fc2a4d7f"} Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.734438 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.917879 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwrwl\" (UniqueName: \"kubernetes.io/projected/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-kube-api-access-cwrwl\") pod \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.917945 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovncontroller-config-0\") pod \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.918053 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-inventory\") pod \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.918076 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovn-combined-ca-bundle\") pod \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.918114 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ssh-key-openstack-edpm-ipam\") pod \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.926036 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-kube-api-access-cwrwl" (OuterVolumeSpecName: "kube-api-access-cwrwl") pod "46cb2975-a7e0-4fe7-8f91-8194638b0c1f" (UID: "46cb2975-a7e0-4fe7-8f91-8194638b0c1f"). InnerVolumeSpecName "kube-api-access-cwrwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.926480 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "46cb2975-a7e0-4fe7-8f91-8194638b0c1f" (UID: "46cb2975-a7e0-4fe7-8f91-8194638b0c1f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.940025 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" event={"ID":"46cb2975-a7e0-4fe7-8f91-8194638b0c1f","Type":"ContainerDied","Data":"903b6991a72997a5a0d659da90a7ce27af0c6725d07ea7083a041a68ae8ca1ed"} Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.940089 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="903b6991a72997a5a0d659da90a7ce27af0c6725d07ea7083a041a68ae8ca1ed" Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.940264 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-zqbsw" Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.956752 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "46cb2975-a7e0-4fe7-8f91-8194638b0c1f" (UID: "46cb2975-a7e0-4fe7-8f91-8194638b0c1f"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:20:34 crc kubenswrapper[4767]: E0317 16:20:34.974088 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-inventory podName:46cb2975-a7e0-4fe7-8f91-8194638b0c1f nodeName:}" failed. No retries permitted until 2026-03-17 16:20:35.47403474 +0000 UTC m=+2626.887350787 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-inventory") pod "46cb2975-a7e0-4fe7-8f91-8194638b0c1f" (UID: "46cb2975-a7e0-4fe7-8f91-8194638b0c1f") : error deleting /var/lib/kubelet/pods/46cb2975-a7e0-4fe7-8f91-8194638b0c1f/volume-subpaths: remove /var/lib/kubelet/pods/46cb2975-a7e0-4fe7-8f91-8194638b0c1f/volume-subpaths: no such file or directory Mar 17 16:20:34 crc kubenswrapper[4767]: I0317 16:20:34.980814 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "46cb2975-a7e0-4fe7-8f91-8194638b0c1f" (UID: "46cb2975-a7e0-4fe7-8f91-8194638b0c1f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.022268 4767 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.022315 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.022327 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwrwl\" (UniqueName: \"kubernetes.io/projected/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-kube-api-access-cwrwl\") on node \"crc\" DevicePath \"\"" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.022337 4767 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.100138 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv"] Mar 17 16:20:35 crc kubenswrapper[4767]: E0317 16:20:35.101187 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8411da78-bfed-47a5-8e0d-1ed245fecd48" containerName="oc" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.101214 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8411da78-bfed-47a5-8e0d-1ed245fecd48" containerName="oc" Mar 17 16:20:35 crc kubenswrapper[4767]: E0317 16:20:35.101249 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerName="extract-content" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.101259 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerName="extract-content" Mar 17 16:20:35 crc kubenswrapper[4767]: E0317 16:20:35.101286 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46cb2975-a7e0-4fe7-8f91-8194638b0c1f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.101296 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="46cb2975-a7e0-4fe7-8f91-8194638b0c1f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 17 16:20:35 crc kubenswrapper[4767]: E0317 16:20:35.101314 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerName="registry-server" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.101324 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerName="registry-server" Mar 17 16:20:35 crc kubenswrapper[4767]: E0317 16:20:35.101350 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerName="extract-utilities" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.101358 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerName="extract-utilities" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.101713 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="8411da78-bfed-47a5-8e0d-1ed245fecd48" containerName="oc" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.101747 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="46cb2975-a7e0-4fe7-8f91-8194638b0c1f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.101768 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af033a0-6d5d-4719-b732-aa72dcd4c883" containerName="registry-server" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.103596 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.110005 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.110290 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.126701 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv"] Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.229423 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.229525 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.229570 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.229591 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.229611 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7899\" (UniqueName: \"kubernetes.io/projected/32f6c2b0-40b1-4802-a41b-dc600626b8ed-kube-api-access-s7899\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.229693 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.332698 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.332774 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.332798 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7899\" (UniqueName: \"kubernetes.io/projected/32f6c2b0-40b1-4802-a41b-dc600626b8ed-kube-api-access-s7899\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.332907 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.333036 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.333096 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.338019 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.338142 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.338414 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.340823 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.350143 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.353154 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7899\" (UniqueName: \"kubernetes.io/projected/32f6c2b0-40b1-4802-a41b-dc600626b8ed-kube-api-access-s7899\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.430611 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.548982 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-inventory\") pod \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\" (UID: \"46cb2975-a7e0-4fe7-8f91-8194638b0c1f\") " Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.554444 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-inventory" (OuterVolumeSpecName: "inventory") pod "46cb2975-a7e0-4fe7-8f91-8194638b0c1f" (UID: "46cb2975-a7e0-4fe7-8f91-8194638b0c1f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:20:35 crc kubenswrapper[4767]: I0317 16:20:35.652648 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46cb2975-a7e0-4fe7-8f91-8194638b0c1f-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:20:36 crc kubenswrapper[4767]: I0317 16:20:36.043928 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv"] Mar 17 16:20:36 crc kubenswrapper[4767]: I0317 16:20:36.966988 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" event={"ID":"32f6c2b0-40b1-4802-a41b-dc600626b8ed","Type":"ContainerStarted","Data":"37db8e7c7c40507bf3a579f6f6da18c56560367d33d401f6356ba98425e9b6ac"} Mar 17 16:20:37 crc kubenswrapper[4767]: I0317 16:20:37.980833 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" event={"ID":"32f6c2b0-40b1-4802-a41b-dc600626b8ed","Type":"ContainerStarted","Data":"03e59ba7e302bbfed37466a1c7cbb3ccbf0f7da57b6ad17b9bb1471b08a285e3"} Mar 17 16:20:38 crc kubenswrapper[4767]: I0317 16:20:38.007294 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" podStartSLOduration=2.417825648 podStartE2EDuration="3.007262497s" podCreationTimestamp="2026-03-17 16:20:35 +0000 UTC" firstStartedPulling="2026-03-17 16:20:36.053992416 +0000 UTC m=+2627.467308463" lastFinishedPulling="2026-03-17 16:20:36.643429265 +0000 UTC m=+2628.056745312" observedRunningTime="2026-03-17 16:20:38.002065346 +0000 UTC m=+2629.415381393" watchObservedRunningTime="2026-03-17 16:20:38.007262497 +0000 UTC m=+2629.420578544" Mar 17 16:20:42 crc kubenswrapper[4767]: I0317 16:20:42.354763 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:20:42 crc kubenswrapper[4767]: E0317 16:20:42.355807 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:20:53 crc kubenswrapper[4767]: I0317 16:20:53.355503 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:20:53 crc kubenswrapper[4767]: E0317 16:20:53.356475 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:21:06 crc kubenswrapper[4767]: I0317 16:21:06.354891 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:21:07 crc kubenswrapper[4767]: I0317 16:21:07.637579 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"96d83a37f6a1880459e8665ce4e3972ef0fe9f08d8096a29a808d28c380d87b1"} Mar 17 16:21:25 crc kubenswrapper[4767]: I0317 16:21:25.024620 4767 generic.go:334] "Generic (PLEG): container finished" podID="32f6c2b0-40b1-4802-a41b-dc600626b8ed" containerID="03e59ba7e302bbfed37466a1c7cbb3ccbf0f7da57b6ad17b9bb1471b08a285e3" exitCode=0 Mar 17 16:21:25 crc kubenswrapper[4767]: I0317 16:21:25.024715 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" event={"ID":"32f6c2b0-40b1-4802-a41b-dc600626b8ed","Type":"ContainerDied","Data":"03e59ba7e302bbfed37466a1c7cbb3ccbf0f7da57b6ad17b9bb1471b08a285e3"} Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.573579 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.734118 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-inventory\") pod \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.734228 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-ovn-metadata-agent-neutron-config-0\") pod \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.734260 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-metadata-combined-ca-bundle\") pod \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.734295 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7899\" (UniqueName: \"kubernetes.io/projected/32f6c2b0-40b1-4802-a41b-dc600626b8ed-kube-api-access-s7899\") pod \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.734379 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-ssh-key-openstack-edpm-ipam\") pod \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.734610 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-nova-metadata-neutron-config-0\") pod \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\" (UID: \"32f6c2b0-40b1-4802-a41b-dc600626b8ed\") " Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.740410 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32f6c2b0-40b1-4802-a41b-dc600626b8ed-kube-api-access-s7899" (OuterVolumeSpecName: "kube-api-access-s7899") pod "32f6c2b0-40b1-4802-a41b-dc600626b8ed" (UID: "32f6c2b0-40b1-4802-a41b-dc600626b8ed"). InnerVolumeSpecName "kube-api-access-s7899". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.741676 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "32f6c2b0-40b1-4802-a41b-dc600626b8ed" (UID: "32f6c2b0-40b1-4802-a41b-dc600626b8ed"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.772048 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "32f6c2b0-40b1-4802-a41b-dc600626b8ed" (UID: "32f6c2b0-40b1-4802-a41b-dc600626b8ed"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.773300 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-inventory" (OuterVolumeSpecName: "inventory") pod "32f6c2b0-40b1-4802-a41b-dc600626b8ed" (UID: "32f6c2b0-40b1-4802-a41b-dc600626b8ed"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.783777 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "32f6c2b0-40b1-4802-a41b-dc600626b8ed" (UID: "32f6c2b0-40b1-4802-a41b-dc600626b8ed"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.792237 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "32f6c2b0-40b1-4802-a41b-dc600626b8ed" (UID: "32f6c2b0-40b1-4802-a41b-dc600626b8ed"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.838732 4767 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.838774 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.838789 4767 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.838801 4767 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.838813 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7899\" (UniqueName: \"kubernetes.io/projected/32f6c2b0-40b1-4802-a41b-dc600626b8ed-kube-api-access-s7899\") on node \"crc\" DevicePath \"\"" Mar 17 16:21:26 crc kubenswrapper[4767]: I0317 16:21:26.838824 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/32f6c2b0-40b1-4802-a41b-dc600626b8ed-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.226077 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" event={"ID":"32f6c2b0-40b1-4802-a41b-dc600626b8ed","Type":"ContainerDied","Data":"37db8e7c7c40507bf3a579f6f6da18c56560367d33d401f6356ba98425e9b6ac"} Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.226488 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37db8e7c7c40507bf3a579f6f6da18c56560367d33d401f6356ba98425e9b6ac" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.226331 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rkkgv" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.410489 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q"] Mar 17 16:21:27 crc kubenswrapper[4767]: E0317 16:21:27.411425 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32f6c2b0-40b1-4802-a41b-dc600626b8ed" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.411455 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="32f6c2b0-40b1-4802-a41b-dc600626b8ed" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.411808 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="32f6c2b0-40b1-4802-a41b-dc600626b8ed" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.412831 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.415650 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.416494 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.416534 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.416658 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.416830 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.441550 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q"] Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.593758 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.593935 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6452\" (UniqueName: \"kubernetes.io/projected/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-kube-api-access-v6452\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.594009 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.594144 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.594208 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.696987 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.697053 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.697167 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.697311 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6452\" (UniqueName: \"kubernetes.io/projected/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-kube-api-access-v6452\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.697390 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.703056 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.703524 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.703549 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.709569 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.716974 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6452\" (UniqueName: \"kubernetes.io/projected/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-kube-api-access-v6452\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:27 crc kubenswrapper[4767]: I0317 16:21:27.739733 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:21:28 crc kubenswrapper[4767]: I0317 16:21:28.407639 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q"] Mar 17 16:21:28 crc kubenswrapper[4767]: W0317 16:21:28.424505 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51c8beb5_c5f2_4b8c_928a_52fa6f08b1ff.slice/crio-239943678ecab6efba920e63cdd25e1b5ae9cb15ed4c5f11fcf25d91beb01156 WatchSource:0}: Error finding container 239943678ecab6efba920e63cdd25e1b5ae9cb15ed4c5f11fcf25d91beb01156: Status 404 returned error can't find the container with id 239943678ecab6efba920e63cdd25e1b5ae9cb15ed4c5f11fcf25d91beb01156 Mar 17 16:21:29 crc kubenswrapper[4767]: I0317 16:21:29.294192 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" event={"ID":"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff","Type":"ContainerStarted","Data":"239943678ecab6efba920e63cdd25e1b5ae9cb15ed4c5f11fcf25d91beb01156"} Mar 17 16:21:30 crc kubenswrapper[4767]: I0317 16:21:30.306965 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" event={"ID":"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff","Type":"ContainerStarted","Data":"312dc3d5dd526bfc1a53accd4c53ca641edc0498170356d4a92351025034fae1"} Mar 17 16:21:30 crc kubenswrapper[4767]: I0317 16:21:30.332661 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" podStartSLOduration=2.67131613 podStartE2EDuration="3.332622029s" podCreationTimestamp="2026-03-17 16:21:27 +0000 UTC" firstStartedPulling="2026-03-17 16:21:28.427538103 +0000 UTC m=+2679.840854150" lastFinishedPulling="2026-03-17 16:21:29.088844002 +0000 UTC m=+2680.502160049" observedRunningTime="2026-03-17 16:21:30.328098348 +0000 UTC m=+2681.741414425" watchObservedRunningTime="2026-03-17 16:21:30.332622029 +0000 UTC m=+2681.745938076" Mar 17 16:22:00 crc kubenswrapper[4767]: I0317 16:22:00.159790 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562742-77vcr"] Mar 17 16:22:00 crc kubenswrapper[4767]: I0317 16:22:00.162849 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562742-77vcr" Mar 17 16:22:00 crc kubenswrapper[4767]: I0317 16:22:00.171994 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:22:00 crc kubenswrapper[4767]: I0317 16:22:00.172064 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:22:00 crc kubenswrapper[4767]: I0317 16:22:00.172556 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:22:00 crc kubenswrapper[4767]: I0317 16:22:00.180926 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562742-77vcr"] Mar 17 16:22:00 crc kubenswrapper[4767]: I0317 16:22:00.283900 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vwdf\" (UniqueName: \"kubernetes.io/projected/00f74843-f644-49ff-8799-2a600a1f17b7-kube-api-access-6vwdf\") pod \"auto-csr-approver-29562742-77vcr\" (UID: \"00f74843-f644-49ff-8799-2a600a1f17b7\") " pod="openshift-infra/auto-csr-approver-29562742-77vcr" Mar 17 16:22:00 crc kubenswrapper[4767]: I0317 16:22:00.387392 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vwdf\" (UniqueName: \"kubernetes.io/projected/00f74843-f644-49ff-8799-2a600a1f17b7-kube-api-access-6vwdf\") pod \"auto-csr-approver-29562742-77vcr\" (UID: \"00f74843-f644-49ff-8799-2a600a1f17b7\") " pod="openshift-infra/auto-csr-approver-29562742-77vcr" Mar 17 16:22:00 crc kubenswrapper[4767]: I0317 16:22:00.422349 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vwdf\" (UniqueName: \"kubernetes.io/projected/00f74843-f644-49ff-8799-2a600a1f17b7-kube-api-access-6vwdf\") pod \"auto-csr-approver-29562742-77vcr\" (UID: \"00f74843-f644-49ff-8799-2a600a1f17b7\") " pod="openshift-infra/auto-csr-approver-29562742-77vcr" Mar 17 16:22:00 crc kubenswrapper[4767]: I0317 16:22:00.490635 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562742-77vcr" Mar 17 16:22:01 crc kubenswrapper[4767]: I0317 16:22:01.176031 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562742-77vcr"] Mar 17 16:22:01 crc kubenswrapper[4767]: W0317 16:22:01.192348 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00f74843_f644_49ff_8799_2a600a1f17b7.slice/crio-366e1fa2cce896aac1aa418e7d1b8a16cc5bc83f33d6e3bf8f1701978dc0c338 WatchSource:0}: Error finding container 366e1fa2cce896aac1aa418e7d1b8a16cc5bc83f33d6e3bf8f1701978dc0c338: Status 404 returned error can't find the container with id 366e1fa2cce896aac1aa418e7d1b8a16cc5bc83f33d6e3bf8f1701978dc0c338 Mar 17 16:22:01 crc kubenswrapper[4767]: I0317 16:22:01.696027 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562742-77vcr" event={"ID":"00f74843-f644-49ff-8799-2a600a1f17b7","Type":"ContainerStarted","Data":"366e1fa2cce896aac1aa418e7d1b8a16cc5bc83f33d6e3bf8f1701978dc0c338"} Mar 17 16:22:02 crc kubenswrapper[4767]: I0317 16:22:02.726607 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562742-77vcr" event={"ID":"00f74843-f644-49ff-8799-2a600a1f17b7","Type":"ContainerStarted","Data":"b743bd5e5200ce703e9582a3c0c4c8c22ca766ee8284bc2d8cf1ccc9b103407c"} Mar 17 16:22:02 crc kubenswrapper[4767]: I0317 16:22:02.749466 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562742-77vcr" podStartSLOduration=1.733925945 podStartE2EDuration="2.749430384s" podCreationTimestamp="2026-03-17 16:22:00 +0000 UTC" firstStartedPulling="2026-03-17 16:22:01.198468717 +0000 UTC m=+2712.611784764" lastFinishedPulling="2026-03-17 16:22:02.213973156 +0000 UTC m=+2713.627289203" observedRunningTime="2026-03-17 16:22:02.740042432 +0000 UTC m=+2714.153358489" watchObservedRunningTime="2026-03-17 16:22:02.749430384 +0000 UTC m=+2714.162746431" Mar 17 16:22:03 crc kubenswrapper[4767]: I0317 16:22:03.741941 4767 generic.go:334] "Generic (PLEG): container finished" podID="00f74843-f644-49ff-8799-2a600a1f17b7" containerID="b743bd5e5200ce703e9582a3c0c4c8c22ca766ee8284bc2d8cf1ccc9b103407c" exitCode=0 Mar 17 16:22:03 crc kubenswrapper[4767]: I0317 16:22:03.742028 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562742-77vcr" event={"ID":"00f74843-f644-49ff-8799-2a600a1f17b7","Type":"ContainerDied","Data":"b743bd5e5200ce703e9582a3c0c4c8c22ca766ee8284bc2d8cf1ccc9b103407c"} Mar 17 16:22:05 crc kubenswrapper[4767]: I0317 16:22:05.190997 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562742-77vcr" Mar 17 16:22:05 crc kubenswrapper[4767]: I0317 16:22:05.259635 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vwdf\" (UniqueName: \"kubernetes.io/projected/00f74843-f644-49ff-8799-2a600a1f17b7-kube-api-access-6vwdf\") pod \"00f74843-f644-49ff-8799-2a600a1f17b7\" (UID: \"00f74843-f644-49ff-8799-2a600a1f17b7\") " Mar 17 16:22:05 crc kubenswrapper[4767]: I0317 16:22:05.269617 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00f74843-f644-49ff-8799-2a600a1f17b7-kube-api-access-6vwdf" (OuterVolumeSpecName: "kube-api-access-6vwdf") pod "00f74843-f644-49ff-8799-2a600a1f17b7" (UID: "00f74843-f644-49ff-8799-2a600a1f17b7"). InnerVolumeSpecName "kube-api-access-6vwdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:22:05 crc kubenswrapper[4767]: I0317 16:22:05.363605 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vwdf\" (UniqueName: \"kubernetes.io/projected/00f74843-f644-49ff-8799-2a600a1f17b7-kube-api-access-6vwdf\") on node \"crc\" DevicePath \"\"" Mar 17 16:22:05 crc kubenswrapper[4767]: I0317 16:22:05.805355 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562742-77vcr" event={"ID":"00f74843-f644-49ff-8799-2a600a1f17b7","Type":"ContainerDied","Data":"366e1fa2cce896aac1aa418e7d1b8a16cc5bc83f33d6e3bf8f1701978dc0c338"} Mar 17 16:22:05 crc kubenswrapper[4767]: I0317 16:22:05.806413 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562742-77vcr" Mar 17 16:22:05 crc kubenswrapper[4767]: I0317 16:22:05.806598 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="366e1fa2cce896aac1aa418e7d1b8a16cc5bc83f33d6e3bf8f1701978dc0c338" Mar 17 16:22:05 crc kubenswrapper[4767]: I0317 16:22:05.874609 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562736-gnl2g"] Mar 17 16:22:05 crc kubenswrapper[4767]: I0317 16:22:05.886656 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562736-gnl2g"] Mar 17 16:22:07 crc kubenswrapper[4767]: I0317 16:22:07.370884 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c54e6dd-f156-4dcb-a753-a3a802735571" path="/var/lib/kubelet/pods/3c54e6dd-f156-4dcb-a753-a3a802735571/volumes" Mar 17 16:22:20 crc kubenswrapper[4767]: I0317 16:22:20.929388 4767 scope.go:117] "RemoveContainer" containerID="2c0ea4042d1e820525c788a4678a5648dde8618947b37195993f7797731cd96e" Mar 17 16:23:34 crc kubenswrapper[4767]: I0317 16:23:34.166461 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:23:34 crc kubenswrapper[4767]: I0317 16:23:34.167089 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.158310 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562744-n9hh5"] Mar 17 16:24:00 crc kubenswrapper[4767]: E0317 16:24:00.159854 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f74843-f644-49ff-8799-2a600a1f17b7" containerName="oc" Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.159875 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f74843-f644-49ff-8799-2a600a1f17b7" containerName="oc" Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.160202 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="00f74843-f644-49ff-8799-2a600a1f17b7" containerName="oc" Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.161595 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562744-n9hh5" Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.164379 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.164570 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.164937 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.177221 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562744-n9hh5"] Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.294418 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcsrk\" (UniqueName: \"kubernetes.io/projected/52c06b1a-98eb-489e-acc4-e51e0bd5bf0e-kube-api-access-gcsrk\") pod \"auto-csr-approver-29562744-n9hh5\" (UID: \"52c06b1a-98eb-489e-acc4-e51e0bd5bf0e\") " pod="openshift-infra/auto-csr-approver-29562744-n9hh5" Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.397778 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcsrk\" (UniqueName: \"kubernetes.io/projected/52c06b1a-98eb-489e-acc4-e51e0bd5bf0e-kube-api-access-gcsrk\") pod \"auto-csr-approver-29562744-n9hh5\" (UID: \"52c06b1a-98eb-489e-acc4-e51e0bd5bf0e\") " pod="openshift-infra/auto-csr-approver-29562744-n9hh5" Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.425100 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcsrk\" (UniqueName: \"kubernetes.io/projected/52c06b1a-98eb-489e-acc4-e51e0bd5bf0e-kube-api-access-gcsrk\") pod \"auto-csr-approver-29562744-n9hh5\" (UID: \"52c06b1a-98eb-489e-acc4-e51e0bd5bf0e\") " pod="openshift-infra/auto-csr-approver-29562744-n9hh5" Mar 17 16:24:00 crc kubenswrapper[4767]: I0317 16:24:00.485746 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562744-n9hh5" Mar 17 16:24:01 crc kubenswrapper[4767]: I0317 16:24:01.002732 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562744-n9hh5"] Mar 17 16:24:01 crc kubenswrapper[4767]: I0317 16:24:01.018281 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 16:24:01 crc kubenswrapper[4767]: I0317 16:24:01.809371 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562744-n9hh5" event={"ID":"52c06b1a-98eb-489e-acc4-e51e0bd5bf0e","Type":"ContainerStarted","Data":"155aaa35bfb0594b9df1a69ba030fb386d0895fdacf3580856dea86175cd37ec"} Mar 17 16:24:02 crc kubenswrapper[4767]: I0317 16:24:02.824732 4767 generic.go:334] "Generic (PLEG): container finished" podID="52c06b1a-98eb-489e-acc4-e51e0bd5bf0e" containerID="60ade6192c4d4cff1c4810b7a9364170d213c36e2b1bcd75d7d39b7196c2dff5" exitCode=0 Mar 17 16:24:02 crc kubenswrapper[4767]: I0317 16:24:02.824960 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562744-n9hh5" event={"ID":"52c06b1a-98eb-489e-acc4-e51e0bd5bf0e","Type":"ContainerDied","Data":"60ade6192c4d4cff1c4810b7a9364170d213c36e2b1bcd75d7d39b7196c2dff5"} Mar 17 16:24:04 crc kubenswrapper[4767]: I0317 16:24:04.166519 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:24:04 crc kubenswrapper[4767]: I0317 16:24:04.167028 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:24:04 crc kubenswrapper[4767]: I0317 16:24:04.339608 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562744-n9hh5" Mar 17 16:24:04 crc kubenswrapper[4767]: I0317 16:24:04.525833 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcsrk\" (UniqueName: \"kubernetes.io/projected/52c06b1a-98eb-489e-acc4-e51e0bd5bf0e-kube-api-access-gcsrk\") pod \"52c06b1a-98eb-489e-acc4-e51e0bd5bf0e\" (UID: \"52c06b1a-98eb-489e-acc4-e51e0bd5bf0e\") " Mar 17 16:24:04 crc kubenswrapper[4767]: I0317 16:24:04.532822 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52c06b1a-98eb-489e-acc4-e51e0bd5bf0e-kube-api-access-gcsrk" (OuterVolumeSpecName: "kube-api-access-gcsrk") pod "52c06b1a-98eb-489e-acc4-e51e0bd5bf0e" (UID: "52c06b1a-98eb-489e-acc4-e51e0bd5bf0e"). InnerVolumeSpecName "kube-api-access-gcsrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:24:04 crc kubenswrapper[4767]: I0317 16:24:04.629812 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcsrk\" (UniqueName: \"kubernetes.io/projected/52c06b1a-98eb-489e-acc4-e51e0bd5bf0e-kube-api-access-gcsrk\") on node \"crc\" DevicePath \"\"" Mar 17 16:24:04 crc kubenswrapper[4767]: I0317 16:24:04.852004 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562744-n9hh5" event={"ID":"52c06b1a-98eb-489e-acc4-e51e0bd5bf0e","Type":"ContainerDied","Data":"155aaa35bfb0594b9df1a69ba030fb386d0895fdacf3580856dea86175cd37ec"} Mar 17 16:24:04 crc kubenswrapper[4767]: I0317 16:24:04.852427 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="155aaa35bfb0594b9df1a69ba030fb386d0895fdacf3580856dea86175cd37ec" Mar 17 16:24:04 crc kubenswrapper[4767]: I0317 16:24:04.852109 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562744-n9hh5" Mar 17 16:24:05 crc kubenswrapper[4767]: I0317 16:24:05.416764 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562738-k77dn"] Mar 17 16:24:05 crc kubenswrapper[4767]: I0317 16:24:05.430064 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562738-k77dn"] Mar 17 16:24:07 crc kubenswrapper[4767]: I0317 16:24:07.465019 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bd93414-e722-485f-a4e5-453ef7d6c4aa" path="/var/lib/kubelet/pods/8bd93414-e722-485f-a4e5-453ef7d6c4aa/volumes" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.211859 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6vcxf"] Mar 17 16:24:12 crc kubenswrapper[4767]: E0317 16:24:12.213289 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52c06b1a-98eb-489e-acc4-e51e0bd5bf0e" containerName="oc" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.213308 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="52c06b1a-98eb-489e-acc4-e51e0bd5bf0e" containerName="oc" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.213570 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="52c06b1a-98eb-489e-acc4-e51e0bd5bf0e" containerName="oc" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.215710 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.227693 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vcxf"] Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.386326 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-catalog-content\") pod \"community-operators-6vcxf\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.386433 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6mlh\" (UniqueName: \"kubernetes.io/projected/62df5734-30cc-474a-933a-ebaf21917d99-kube-api-access-k6mlh\") pod \"community-operators-6vcxf\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.386466 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-utilities\") pod \"community-operators-6vcxf\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.489491 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-catalog-content\") pod \"community-operators-6vcxf\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.490007 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6mlh\" (UniqueName: \"kubernetes.io/projected/62df5734-30cc-474a-933a-ebaf21917d99-kube-api-access-k6mlh\") pod \"community-operators-6vcxf\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.490056 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-utilities\") pod \"community-operators-6vcxf\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.491505 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-catalog-content\") pod \"community-operators-6vcxf\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.491668 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-utilities\") pod \"community-operators-6vcxf\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.516479 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6mlh\" (UniqueName: \"kubernetes.io/projected/62df5734-30cc-474a-933a-ebaf21917d99-kube-api-access-k6mlh\") pod \"community-operators-6vcxf\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:12 crc kubenswrapper[4767]: I0317 16:24:12.539385 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:13 crc kubenswrapper[4767]: I0317 16:24:13.303144 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vcxf"] Mar 17 16:24:13 crc kubenswrapper[4767]: W0317 16:24:13.323294 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62df5734_30cc_474a_933a_ebaf21917d99.slice/crio-2f511e9be477b5abd1da1f725449c1dd3ff244d043a4f7a20b8c09af247f3054 WatchSource:0}: Error finding container 2f511e9be477b5abd1da1f725449c1dd3ff244d043a4f7a20b8c09af247f3054: Status 404 returned error can't find the container with id 2f511e9be477b5abd1da1f725449c1dd3ff244d043a4f7a20b8c09af247f3054 Mar 17 16:24:13 crc kubenswrapper[4767]: I0317 16:24:13.983105 4767 generic.go:334] "Generic (PLEG): container finished" podID="62df5734-30cc-474a-933a-ebaf21917d99" containerID="242d7e5d52625db9b88b4ec5a7326cd7cd028eaeb12b433c8af8b1ef0762bbea" exitCode=0 Mar 17 16:24:13 crc kubenswrapper[4767]: I0317 16:24:13.983234 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vcxf" event={"ID":"62df5734-30cc-474a-933a-ebaf21917d99","Type":"ContainerDied","Data":"242d7e5d52625db9b88b4ec5a7326cd7cd028eaeb12b433c8af8b1ef0762bbea"} Mar 17 16:24:13 crc kubenswrapper[4767]: I0317 16:24:13.983433 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vcxf" event={"ID":"62df5734-30cc-474a-933a-ebaf21917d99","Type":"ContainerStarted","Data":"2f511e9be477b5abd1da1f725449c1dd3ff244d043a4f7a20b8c09af247f3054"} Mar 17 16:24:14 crc kubenswrapper[4767]: I0317 16:24:14.998118 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vcxf" event={"ID":"62df5734-30cc-474a-933a-ebaf21917d99","Type":"ContainerStarted","Data":"a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d"} Mar 17 16:24:17 crc kubenswrapper[4767]: I0317 16:24:17.104992 4767 generic.go:334] "Generic (PLEG): container finished" podID="62df5734-30cc-474a-933a-ebaf21917d99" containerID="a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d" exitCode=0 Mar 17 16:24:17 crc kubenswrapper[4767]: I0317 16:24:17.105098 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vcxf" event={"ID":"62df5734-30cc-474a-933a-ebaf21917d99","Type":"ContainerDied","Data":"a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d"} Mar 17 16:24:19 crc kubenswrapper[4767]: I0317 16:24:19.136618 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vcxf" event={"ID":"62df5734-30cc-474a-933a-ebaf21917d99","Type":"ContainerStarted","Data":"3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269"} Mar 17 16:24:19 crc kubenswrapper[4767]: I0317 16:24:19.182969 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6vcxf" podStartSLOduration=2.558184771 podStartE2EDuration="7.182926811s" podCreationTimestamp="2026-03-17 16:24:12 +0000 UTC" firstStartedPulling="2026-03-17 16:24:13.986047652 +0000 UTC m=+2845.399363699" lastFinishedPulling="2026-03-17 16:24:18.610789692 +0000 UTC m=+2850.024105739" observedRunningTime="2026-03-17 16:24:19.175692267 +0000 UTC m=+2850.589008334" watchObservedRunningTime="2026-03-17 16:24:19.182926811 +0000 UTC m=+2850.596242858" Mar 17 16:24:21 crc kubenswrapper[4767]: I0317 16:24:21.071285 4767 scope.go:117] "RemoveContainer" containerID="2981b86328461469ddfe474c26d47ddd87adcb49a171fd4a7b97b76b6f085f36" Mar 17 16:24:22 crc kubenswrapper[4767]: I0317 16:24:22.540619 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:22 crc kubenswrapper[4767]: I0317 16:24:22.541289 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:22 crc kubenswrapper[4767]: I0317 16:24:22.600732 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:23 crc kubenswrapper[4767]: I0317 16:24:23.262427 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:23 crc kubenswrapper[4767]: I0317 16:24:23.318495 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6vcxf"] Mar 17 16:24:25 crc kubenswrapper[4767]: I0317 16:24:25.231485 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6vcxf" podUID="62df5734-30cc-474a-933a-ebaf21917d99" containerName="registry-server" containerID="cri-o://3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269" gracePeriod=2 Mar 17 16:24:26 crc kubenswrapper[4767]: I0317 16:24:26.894979 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.046721 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-utilities\") pod \"62df5734-30cc-474a-933a-ebaf21917d99\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.046998 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6mlh\" (UniqueName: \"kubernetes.io/projected/62df5734-30cc-474a-933a-ebaf21917d99-kube-api-access-k6mlh\") pod \"62df5734-30cc-474a-933a-ebaf21917d99\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.047089 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-catalog-content\") pod \"62df5734-30cc-474a-933a-ebaf21917d99\" (UID: \"62df5734-30cc-474a-933a-ebaf21917d99\") " Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.048204 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-utilities" (OuterVolumeSpecName: "utilities") pod "62df5734-30cc-474a-933a-ebaf21917d99" (UID: "62df5734-30cc-474a-933a-ebaf21917d99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.104478 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62df5734-30cc-474a-933a-ebaf21917d99" (UID: "62df5734-30cc-474a-933a-ebaf21917d99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.151305 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.151338 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62df5734-30cc-474a-933a-ebaf21917d99-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.196453 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62df5734-30cc-474a-933a-ebaf21917d99-kube-api-access-k6mlh" (OuterVolumeSpecName: "kube-api-access-k6mlh") pod "62df5734-30cc-474a-933a-ebaf21917d99" (UID: "62df5734-30cc-474a-933a-ebaf21917d99"). InnerVolumeSpecName "kube-api-access-k6mlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.254498 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6mlh\" (UniqueName: \"kubernetes.io/projected/62df5734-30cc-474a-933a-ebaf21917d99-kube-api-access-k6mlh\") on node \"crc\" DevicePath \"\"" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.278323 4767 generic.go:334] "Generic (PLEG): container finished" podID="62df5734-30cc-474a-933a-ebaf21917d99" containerID="3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269" exitCode=0 Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.278384 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vcxf" event={"ID":"62df5734-30cc-474a-933a-ebaf21917d99","Type":"ContainerDied","Data":"3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269"} Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.278400 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vcxf" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.278435 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vcxf" event={"ID":"62df5734-30cc-474a-933a-ebaf21917d99","Type":"ContainerDied","Data":"2f511e9be477b5abd1da1f725449c1dd3ff244d043a4f7a20b8c09af247f3054"} Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.278462 4767 scope.go:117] "RemoveContainer" containerID="3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.311732 4767 scope.go:117] "RemoveContainer" containerID="a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.322212 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6vcxf"] Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.333066 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6vcxf"] Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.343264 4767 scope.go:117] "RemoveContainer" containerID="242d7e5d52625db9b88b4ec5a7326cd7cd028eaeb12b433c8af8b1ef0762bbea" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.379911 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62df5734-30cc-474a-933a-ebaf21917d99" path="/var/lib/kubelet/pods/62df5734-30cc-474a-933a-ebaf21917d99/volumes" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.434374 4767 scope.go:117] "RemoveContainer" containerID="3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269" Mar 17 16:24:27 crc kubenswrapper[4767]: E0317 16:24:27.434915 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269\": container with ID starting with 3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269 not found: ID does not exist" containerID="3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.434956 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269"} err="failed to get container status \"3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269\": rpc error: code = NotFound desc = could not find container \"3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269\": container with ID starting with 3393a9c1ba15bcfa045871879ac27d59236ffc91fbc989511a4f1273d6cd7269 not found: ID does not exist" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.434991 4767 scope.go:117] "RemoveContainer" containerID="a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d" Mar 17 16:24:27 crc kubenswrapper[4767]: E0317 16:24:27.435297 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d\": container with ID starting with a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d not found: ID does not exist" containerID="a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.435412 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d"} err="failed to get container status \"a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d\": rpc error: code = NotFound desc = could not find container \"a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d\": container with ID starting with a6112186a51f66e11cba9424c76bbfa1aed4a2da539e0953c0554763e4f1977d not found: ID does not exist" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.435495 4767 scope.go:117] "RemoveContainer" containerID="242d7e5d52625db9b88b4ec5a7326cd7cd028eaeb12b433c8af8b1ef0762bbea" Mar 17 16:24:27 crc kubenswrapper[4767]: E0317 16:24:27.435773 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242d7e5d52625db9b88b4ec5a7326cd7cd028eaeb12b433c8af8b1ef0762bbea\": container with ID starting with 242d7e5d52625db9b88b4ec5a7326cd7cd028eaeb12b433c8af8b1ef0762bbea not found: ID does not exist" containerID="242d7e5d52625db9b88b4ec5a7326cd7cd028eaeb12b433c8af8b1ef0762bbea" Mar 17 16:24:27 crc kubenswrapper[4767]: I0317 16:24:27.435801 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242d7e5d52625db9b88b4ec5a7326cd7cd028eaeb12b433c8af8b1ef0762bbea"} err="failed to get container status \"242d7e5d52625db9b88b4ec5a7326cd7cd028eaeb12b433c8af8b1ef0762bbea\": rpc error: code = NotFound desc = could not find container \"242d7e5d52625db9b88b4ec5a7326cd7cd028eaeb12b433c8af8b1ef0762bbea\": container with ID starting with 242d7e5d52625db9b88b4ec5a7326cd7cd028eaeb12b433c8af8b1ef0762bbea not found: ID does not exist" Mar 17 16:24:34 crc kubenswrapper[4767]: I0317 16:24:34.344464 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:24:34 crc kubenswrapper[4767]: I0317 16:24:34.345133 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:24:34 crc kubenswrapper[4767]: I0317 16:24:34.348045 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 16:24:34 crc kubenswrapper[4767]: I0317 16:24:34.349661 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"96d83a37f6a1880459e8665ce4e3972ef0fe9f08d8096a29a808d28c380d87b1"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 16:24:34 crc kubenswrapper[4767]: I0317 16:24:34.349748 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://96d83a37f6a1880459e8665ce4e3972ef0fe9f08d8096a29a808d28c380d87b1" gracePeriod=600 Mar 17 16:24:35 crc kubenswrapper[4767]: I0317 16:24:35.416932 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="96d83a37f6a1880459e8665ce4e3972ef0fe9f08d8096a29a808d28c380d87b1" exitCode=0 Mar 17 16:24:35 crc kubenswrapper[4767]: I0317 16:24:35.416997 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"96d83a37f6a1880459e8665ce4e3972ef0fe9f08d8096a29a808d28c380d87b1"} Mar 17 16:24:35 crc kubenswrapper[4767]: I0317 16:24:35.418278 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad"} Mar 17 16:24:35 crc kubenswrapper[4767]: I0317 16:24:35.418304 4767 scope.go:117] "RemoveContainer" containerID="0f9bd320bc4b5c640c441c11eb4d95202a22b3873c7988fcbf72fb23424a6f36" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.544940 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l7lfk"] Mar 17 16:24:44 crc kubenswrapper[4767]: E0317 16:24:44.546463 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62df5734-30cc-474a-933a-ebaf21917d99" containerName="registry-server" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.546484 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="62df5734-30cc-474a-933a-ebaf21917d99" containerName="registry-server" Mar 17 16:24:44 crc kubenswrapper[4767]: E0317 16:24:44.546533 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62df5734-30cc-474a-933a-ebaf21917d99" containerName="extract-utilities" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.546542 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="62df5734-30cc-474a-933a-ebaf21917d99" containerName="extract-utilities" Mar 17 16:24:44 crc kubenswrapper[4767]: E0317 16:24:44.546598 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62df5734-30cc-474a-933a-ebaf21917d99" containerName="extract-content" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.546608 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="62df5734-30cc-474a-933a-ebaf21917d99" containerName="extract-content" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.546891 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="62df5734-30cc-474a-933a-ebaf21917d99" containerName="registry-server" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.549350 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.563130 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7lfk"] Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.676848 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-catalog-content\") pod \"redhat-marketplace-l7lfk\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.677031 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-utilities\") pod \"redhat-marketplace-l7lfk\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.677249 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46gpw\" (UniqueName: \"kubernetes.io/projected/87c8b037-e656-4380-afe7-cde338017725-kube-api-access-46gpw\") pod \"redhat-marketplace-l7lfk\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.779963 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46gpw\" (UniqueName: \"kubernetes.io/projected/87c8b037-e656-4380-afe7-cde338017725-kube-api-access-46gpw\") pod \"redhat-marketplace-l7lfk\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.780061 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-catalog-content\") pod \"redhat-marketplace-l7lfk\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.780185 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-utilities\") pod \"redhat-marketplace-l7lfk\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.780710 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-catalog-content\") pod \"redhat-marketplace-l7lfk\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.784436 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-utilities\") pod \"redhat-marketplace-l7lfk\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.811122 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46gpw\" (UniqueName: \"kubernetes.io/projected/87c8b037-e656-4380-afe7-cde338017725-kube-api-access-46gpw\") pod \"redhat-marketplace-l7lfk\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:44 crc kubenswrapper[4767]: I0317 16:24:44.877331 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:45 crc kubenswrapper[4767]: W0317 16:24:45.487208 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87c8b037_e656_4380_afe7_cde338017725.slice/crio-35dece9c41ab4c7ea836392c086b1e9eeb0d67ebcafdb95b18473875fafde04d WatchSource:0}: Error finding container 35dece9c41ab4c7ea836392c086b1e9eeb0d67ebcafdb95b18473875fafde04d: Status 404 returned error can't find the container with id 35dece9c41ab4c7ea836392c086b1e9eeb0d67ebcafdb95b18473875fafde04d Mar 17 16:24:45 crc kubenswrapper[4767]: I0317 16:24:45.487638 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7lfk"] Mar 17 16:24:45 crc kubenswrapper[4767]: I0317 16:24:45.601006 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7lfk" event={"ID":"87c8b037-e656-4380-afe7-cde338017725","Type":"ContainerStarted","Data":"35dece9c41ab4c7ea836392c086b1e9eeb0d67ebcafdb95b18473875fafde04d"} Mar 17 16:24:46 crc kubenswrapper[4767]: I0317 16:24:46.613881 4767 generic.go:334] "Generic (PLEG): container finished" podID="87c8b037-e656-4380-afe7-cde338017725" containerID="35532e9f02cca972852997b72acf292928b6eb13f24c233e1317f640e58b3a95" exitCode=0 Mar 17 16:24:46 crc kubenswrapper[4767]: I0317 16:24:46.613930 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7lfk" event={"ID":"87c8b037-e656-4380-afe7-cde338017725","Type":"ContainerDied","Data":"35532e9f02cca972852997b72acf292928b6eb13f24c233e1317f640e58b3a95"} Mar 17 16:24:48 crc kubenswrapper[4767]: I0317 16:24:48.640372 4767 generic.go:334] "Generic (PLEG): container finished" podID="87c8b037-e656-4380-afe7-cde338017725" containerID="759dcce566c929d897531bfea8dfd5accafb642c28b6ce5cae357c5ffb99ed5a" exitCode=0 Mar 17 16:24:48 crc kubenswrapper[4767]: I0317 16:24:48.640468 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7lfk" event={"ID":"87c8b037-e656-4380-afe7-cde338017725","Type":"ContainerDied","Data":"759dcce566c929d897531bfea8dfd5accafb642c28b6ce5cae357c5ffb99ed5a"} Mar 17 16:24:49 crc kubenswrapper[4767]: I0317 16:24:49.657862 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7lfk" event={"ID":"87c8b037-e656-4380-afe7-cde338017725","Type":"ContainerStarted","Data":"803ca92a35d96ae50a6ae5808bd3ffc418d8c9152afa7fef6c9886594d22e83d"} Mar 17 16:24:49 crc kubenswrapper[4767]: I0317 16:24:49.697620 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l7lfk" podStartSLOduration=3.284031866 podStartE2EDuration="5.69759377s" podCreationTimestamp="2026-03-17 16:24:44 +0000 UTC" firstStartedPulling="2026-03-17 16:24:46.6167028 +0000 UTC m=+2878.030018847" lastFinishedPulling="2026-03-17 16:24:49.030264704 +0000 UTC m=+2880.443580751" observedRunningTime="2026-03-17 16:24:49.68408315 +0000 UTC m=+2881.097399207" watchObservedRunningTime="2026-03-17 16:24:49.69759377 +0000 UTC m=+2881.110909817" Mar 17 16:24:54 crc kubenswrapper[4767]: I0317 16:24:54.878255 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:54 crc kubenswrapper[4767]: I0317 16:24:54.878874 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:54 crc kubenswrapper[4767]: I0317 16:24:54.929350 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:55 crc kubenswrapper[4767]: I0317 16:24:55.800685 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:24:55 crc kubenswrapper[4767]: I0317 16:24:55.862053 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7lfk"] Mar 17 16:24:57 crc kubenswrapper[4767]: I0317 16:24:57.945725 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l7lfk" podUID="87c8b037-e656-4380-afe7-cde338017725" containerName="registry-server" containerID="cri-o://803ca92a35d96ae50a6ae5808bd3ffc418d8c9152afa7fef6c9886594d22e83d" gracePeriod=2 Mar 17 16:24:59 crc kubenswrapper[4767]: I0317 16:24:59.070090 4767 generic.go:334] "Generic (PLEG): container finished" podID="87c8b037-e656-4380-afe7-cde338017725" containerID="803ca92a35d96ae50a6ae5808bd3ffc418d8c9152afa7fef6c9886594d22e83d" exitCode=0 Mar 17 16:24:59 crc kubenswrapper[4767]: I0317 16:24:59.070418 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7lfk" event={"ID":"87c8b037-e656-4380-afe7-cde338017725","Type":"ContainerDied","Data":"803ca92a35d96ae50a6ae5808bd3ffc418d8c9152afa7fef6c9886594d22e83d"} Mar 17 16:25:00 crc kubenswrapper[4767]: I0317 16:25:00.307988 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:25:00 crc kubenswrapper[4767]: I0317 16:25:00.423196 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-utilities\") pod \"87c8b037-e656-4380-afe7-cde338017725\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " Mar 17 16:25:00 crc kubenswrapper[4767]: I0317 16:25:00.423406 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46gpw\" (UniqueName: \"kubernetes.io/projected/87c8b037-e656-4380-afe7-cde338017725-kube-api-access-46gpw\") pod \"87c8b037-e656-4380-afe7-cde338017725\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " Mar 17 16:25:00 crc kubenswrapper[4767]: I0317 16:25:00.423529 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-catalog-content\") pod \"87c8b037-e656-4380-afe7-cde338017725\" (UID: \"87c8b037-e656-4380-afe7-cde338017725\") " Mar 17 16:25:00 crc kubenswrapper[4767]: I0317 16:25:00.424515 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-utilities" (OuterVolumeSpecName: "utilities") pod "87c8b037-e656-4380-afe7-cde338017725" (UID: "87c8b037-e656-4380-afe7-cde338017725"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:25:00 crc kubenswrapper[4767]: I0317 16:25:00.429852 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87c8b037-e656-4380-afe7-cde338017725-kube-api-access-46gpw" (OuterVolumeSpecName: "kube-api-access-46gpw") pod "87c8b037-e656-4380-afe7-cde338017725" (UID: "87c8b037-e656-4380-afe7-cde338017725"). InnerVolumeSpecName "kube-api-access-46gpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:25:00 crc kubenswrapper[4767]: I0317 16:25:00.452662 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87c8b037-e656-4380-afe7-cde338017725" (UID: "87c8b037-e656-4380-afe7-cde338017725"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:25:00 crc kubenswrapper[4767]: I0317 16:25:00.527156 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:25:00 crc kubenswrapper[4767]: I0317 16:25:00.527237 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46gpw\" (UniqueName: \"kubernetes.io/projected/87c8b037-e656-4380-afe7-cde338017725-kube-api-access-46gpw\") on node \"crc\" DevicePath \"\"" Mar 17 16:25:00 crc kubenswrapper[4767]: I0317 16:25:00.527255 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87c8b037-e656-4380-afe7-cde338017725-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:25:01 crc kubenswrapper[4767]: I0317 16:25:01.097793 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7lfk" event={"ID":"87c8b037-e656-4380-afe7-cde338017725","Type":"ContainerDied","Data":"35dece9c41ab4c7ea836392c086b1e9eeb0d67ebcafdb95b18473875fafde04d"} Mar 17 16:25:01 crc kubenswrapper[4767]: I0317 16:25:01.098211 4767 scope.go:117] "RemoveContainer" containerID="803ca92a35d96ae50a6ae5808bd3ffc418d8c9152afa7fef6c9886594d22e83d" Mar 17 16:25:01 crc kubenswrapper[4767]: I0317 16:25:01.097895 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7lfk" Mar 17 16:25:01 crc kubenswrapper[4767]: I0317 16:25:01.137647 4767 scope.go:117] "RemoveContainer" containerID="759dcce566c929d897531bfea8dfd5accafb642c28b6ce5cae357c5ffb99ed5a" Mar 17 16:25:01 crc kubenswrapper[4767]: I0317 16:25:01.145365 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7lfk"] Mar 17 16:25:01 crc kubenswrapper[4767]: I0317 16:25:01.163235 4767 scope.go:117] "RemoveContainer" containerID="35532e9f02cca972852997b72acf292928b6eb13f24c233e1317f640e58b3a95" Mar 17 16:25:01 crc kubenswrapper[4767]: I0317 16:25:01.167234 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7lfk"] Mar 17 16:25:02 crc kubenswrapper[4767]: I0317 16:25:02.175246 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87c8b037-e656-4380-afe7-cde338017725" path="/var/lib/kubelet/pods/87c8b037-e656-4380-afe7-cde338017725/volumes" Mar 17 16:25:43 crc kubenswrapper[4767]: I0317 16:25:43.667606 4767 generic.go:334] "Generic (PLEG): container finished" podID="51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff" containerID="312dc3d5dd526bfc1a53accd4c53ca641edc0498170356d4a92351025034fae1" exitCode=0 Mar 17 16:25:43 crc kubenswrapper[4767]: I0317 16:25:43.667683 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" event={"ID":"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff","Type":"ContainerDied","Data":"312dc3d5dd526bfc1a53accd4c53ca641edc0498170356d4a92351025034fae1"} Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.235787 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.298509 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-ssh-key-openstack-edpm-ipam\") pod \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.298616 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6452\" (UniqueName: \"kubernetes.io/projected/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-kube-api-access-v6452\") pod \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.298647 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-inventory\") pod \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.298693 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-combined-ca-bundle\") pod \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.306473 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-kube-api-access-v6452" (OuterVolumeSpecName: "kube-api-access-v6452") pod "51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff" (UID: "51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff"). InnerVolumeSpecName "kube-api-access-v6452". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.309363 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff" (UID: "51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.339750 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-inventory" (OuterVolumeSpecName: "inventory") pod "51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff" (UID: "51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.345416 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff" (UID: "51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.400564 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-secret-0\") pod \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\" (UID: \"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff\") " Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.402963 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.403162 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6452\" (UniqueName: \"kubernetes.io/projected/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-kube-api-access-v6452\") on node \"crc\" DevicePath \"\"" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.403250 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.403348 4767 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.441201 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff" (UID: "51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.506654 4767 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.699835 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" event={"ID":"51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff","Type":"ContainerDied","Data":"239943678ecab6efba920e63cdd25e1b5ae9cb15ed4c5f11fcf25d91beb01156"} Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.700233 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="239943678ecab6efba920e63cdd25e1b5ae9cb15ed4c5f11fcf25d91beb01156" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.699913 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8nj4q" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.838924 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7"] Mar 17 16:25:45 crc kubenswrapper[4767]: E0317 16:25:45.839531 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c8b037-e656-4380-afe7-cde338017725" containerName="extract-content" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.839570 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c8b037-e656-4380-afe7-cde338017725" containerName="extract-content" Mar 17 16:25:45 crc kubenswrapper[4767]: E0317 16:25:45.839585 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.839593 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 17 16:25:45 crc kubenswrapper[4767]: E0317 16:25:45.839611 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c8b037-e656-4380-afe7-cde338017725" containerName="extract-utilities" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.839618 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c8b037-e656-4380-afe7-cde338017725" containerName="extract-utilities" Mar 17 16:25:45 crc kubenswrapper[4767]: E0317 16:25:45.839636 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c8b037-e656-4380-afe7-cde338017725" containerName="registry-server" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.839645 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c8b037-e656-4380-afe7-cde338017725" containerName="registry-server" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.839956 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="51c8beb5-c5f2-4b8c-928a-52fa6f08b1ff" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.839990 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c8b037-e656-4380-afe7-cde338017725" containerName="registry-server" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.840961 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.845162 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.845276 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.845359 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.845707 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.845976 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.849521 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.849579 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:25:45 crc kubenswrapper[4767]: I0317 16:25:45.867368 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7"] Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.019659 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.019711 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.019804 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.019826 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.019862 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.019907 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.019955 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.020006 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.020055 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx8d8\" (UniqueName: \"kubernetes.io/projected/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-kube-api-access-hx8d8\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.020074 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.020097 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.122839 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.122949 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.123036 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.123123 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.124197 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.125791 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx8d8\" (UniqueName: \"kubernetes.io/projected/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-kube-api-access-hx8d8\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.125842 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.125879 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.125967 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.126019 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.126128 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.126187 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.128400 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.128507 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.129520 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.130745 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.130810 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.131380 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.132026 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.132468 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.132769 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.146112 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx8d8\" (UniqueName: \"kubernetes.io/projected/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-kube-api-access-hx8d8\") pod \"nova-edpm-deployment-openstack-edpm-ipam-9rvm7\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.170105 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:25:46 crc kubenswrapper[4767]: I0317 16:25:46.806130 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7"] Mar 17 16:25:47 crc kubenswrapper[4767]: I0317 16:25:47.729748 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" event={"ID":"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a","Type":"ContainerStarted","Data":"f3cebea0ca8abc3cd021683c0a5439a9cf8feb2cca0f69bad2837e0365ecd30a"} Mar 17 16:25:48 crc kubenswrapper[4767]: I0317 16:25:48.743709 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" event={"ID":"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a","Type":"ContainerStarted","Data":"876a640a3ef3261e4ff5c78ee072b8d66a3731829131a4e6fe0e14b067caca64"} Mar 17 16:25:48 crc kubenswrapper[4767]: I0317 16:25:48.771110 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" podStartSLOduration=3.119288846 podStartE2EDuration="3.771080022s" podCreationTimestamp="2026-03-17 16:25:45 +0000 UTC" firstStartedPulling="2026-03-17 16:25:46.818633702 +0000 UTC m=+2938.231949749" lastFinishedPulling="2026-03-17 16:25:47.470424878 +0000 UTC m=+2938.883740925" observedRunningTime="2026-03-17 16:25:48.764483457 +0000 UTC m=+2940.177799514" watchObservedRunningTime="2026-03-17 16:25:48.771080022 +0000 UTC m=+2940.184396069" Mar 17 16:26:00 crc kubenswrapper[4767]: I0317 16:26:00.161088 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562746-d8mln"] Mar 17 16:26:00 crc kubenswrapper[4767]: I0317 16:26:00.164836 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562746-d8mln" Mar 17 16:26:00 crc kubenswrapper[4767]: I0317 16:26:00.168122 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:26:00 crc kubenswrapper[4767]: I0317 16:26:00.168785 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:26:00 crc kubenswrapper[4767]: I0317 16:26:00.168993 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:26:00 crc kubenswrapper[4767]: I0317 16:26:00.175785 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562746-d8mln"] Mar 17 16:26:00 crc kubenswrapper[4767]: I0317 16:26:00.251683 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmgs6\" (UniqueName: \"kubernetes.io/projected/24066b25-22e0-4da5-add8-ffade9245cb1-kube-api-access-jmgs6\") pod \"auto-csr-approver-29562746-d8mln\" (UID: \"24066b25-22e0-4da5-add8-ffade9245cb1\") " pod="openshift-infra/auto-csr-approver-29562746-d8mln" Mar 17 16:26:00 crc kubenswrapper[4767]: I0317 16:26:00.355476 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmgs6\" (UniqueName: \"kubernetes.io/projected/24066b25-22e0-4da5-add8-ffade9245cb1-kube-api-access-jmgs6\") pod \"auto-csr-approver-29562746-d8mln\" (UID: \"24066b25-22e0-4da5-add8-ffade9245cb1\") " pod="openshift-infra/auto-csr-approver-29562746-d8mln" Mar 17 16:26:00 crc kubenswrapper[4767]: I0317 16:26:00.382387 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmgs6\" (UniqueName: \"kubernetes.io/projected/24066b25-22e0-4da5-add8-ffade9245cb1-kube-api-access-jmgs6\") pod \"auto-csr-approver-29562746-d8mln\" (UID: \"24066b25-22e0-4da5-add8-ffade9245cb1\") " pod="openshift-infra/auto-csr-approver-29562746-d8mln" Mar 17 16:26:00 crc kubenswrapper[4767]: I0317 16:26:00.492896 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562746-d8mln" Mar 17 16:26:01 crc kubenswrapper[4767]: I0317 16:26:01.053694 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562746-d8mln"] Mar 17 16:26:01 crc kubenswrapper[4767]: I0317 16:26:01.923085 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562746-d8mln" event={"ID":"24066b25-22e0-4da5-add8-ffade9245cb1","Type":"ContainerStarted","Data":"29646e60b9efbf4e5d197e1d65b3f3671980ca497094b4a16e1b516bea21605f"} Mar 17 16:26:02 crc kubenswrapper[4767]: I0317 16:26:02.951155 4767 generic.go:334] "Generic (PLEG): container finished" podID="24066b25-22e0-4da5-add8-ffade9245cb1" containerID="1d62129ba6a2096431bfd0c6413558229b0461aec5c0b9cb4a753612823c4433" exitCode=0 Mar 17 16:26:02 crc kubenswrapper[4767]: I0317 16:26:02.951316 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562746-d8mln" event={"ID":"24066b25-22e0-4da5-add8-ffade9245cb1","Type":"ContainerDied","Data":"1d62129ba6a2096431bfd0c6413558229b0461aec5c0b9cb4a753612823c4433"} Mar 17 16:26:04 crc kubenswrapper[4767]: I0317 16:26:04.452877 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562746-d8mln" Mar 17 16:26:04 crc kubenswrapper[4767]: I0317 16:26:04.596069 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmgs6\" (UniqueName: \"kubernetes.io/projected/24066b25-22e0-4da5-add8-ffade9245cb1-kube-api-access-jmgs6\") pod \"24066b25-22e0-4da5-add8-ffade9245cb1\" (UID: \"24066b25-22e0-4da5-add8-ffade9245cb1\") " Mar 17 16:26:04 crc kubenswrapper[4767]: I0317 16:26:04.608389 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24066b25-22e0-4da5-add8-ffade9245cb1-kube-api-access-jmgs6" (OuterVolumeSpecName: "kube-api-access-jmgs6") pod "24066b25-22e0-4da5-add8-ffade9245cb1" (UID: "24066b25-22e0-4da5-add8-ffade9245cb1"). InnerVolumeSpecName "kube-api-access-jmgs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:26:04 crc kubenswrapper[4767]: I0317 16:26:04.700626 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmgs6\" (UniqueName: \"kubernetes.io/projected/24066b25-22e0-4da5-add8-ffade9245cb1-kube-api-access-jmgs6\") on node \"crc\" DevicePath \"\"" Mar 17 16:26:04 crc kubenswrapper[4767]: I0317 16:26:04.978612 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562746-d8mln" event={"ID":"24066b25-22e0-4da5-add8-ffade9245cb1","Type":"ContainerDied","Data":"29646e60b9efbf4e5d197e1d65b3f3671980ca497094b4a16e1b516bea21605f"} Mar 17 16:26:04 crc kubenswrapper[4767]: I0317 16:26:04.978664 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29646e60b9efbf4e5d197e1d65b3f3671980ca497094b4a16e1b516bea21605f" Mar 17 16:26:04 crc kubenswrapper[4767]: I0317 16:26:04.978675 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562746-d8mln" Mar 17 16:26:05 crc kubenswrapper[4767]: I0317 16:26:05.540445 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562740-cv9nj"] Mar 17 16:26:05 crc kubenswrapper[4767]: I0317 16:26:05.553726 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562740-cv9nj"] Mar 17 16:26:07 crc kubenswrapper[4767]: I0317 16:26:07.369555 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8411da78-bfed-47a5-8e0d-1ed245fecd48" path="/var/lib/kubelet/pods/8411da78-bfed-47a5-8e0d-1ed245fecd48/volumes" Mar 17 16:26:21 crc kubenswrapper[4767]: I0317 16:26:21.319019 4767 scope.go:117] "RemoveContainer" containerID="2ace0c423c7d66ab119f54c12faa314382cdc010e67e852f7367de8d77c9cab1" Mar 17 16:26:34 crc kubenswrapper[4767]: I0317 16:26:34.167271 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:26:34 crc kubenswrapper[4767]: I0317 16:26:34.167786 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:27:04 crc kubenswrapper[4767]: I0317 16:27:04.166196 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:27:04 crc kubenswrapper[4767]: I0317 16:27:04.167715 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:27:34 crc kubenswrapper[4767]: I0317 16:27:34.166868 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:27:34 crc kubenswrapper[4767]: I0317 16:27:34.167567 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:27:34 crc kubenswrapper[4767]: I0317 16:27:34.167650 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 16:27:34 crc kubenswrapper[4767]: I0317 16:27:34.168956 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 16:27:34 crc kubenswrapper[4767]: I0317 16:27:34.169031 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" gracePeriod=600 Mar 17 16:27:34 crc kubenswrapper[4767]: E0317 16:27:34.290076 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:27:34 crc kubenswrapper[4767]: I0317 16:27:34.480680 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" exitCode=0 Mar 17 16:27:34 crc kubenswrapper[4767]: I0317 16:27:34.480787 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad"} Mar 17 16:27:34 crc kubenswrapper[4767]: I0317 16:27:34.481079 4767 scope.go:117] "RemoveContainer" containerID="96d83a37f6a1880459e8665ce4e3972ef0fe9f08d8096a29a808d28c380d87b1" Mar 17 16:27:34 crc kubenswrapper[4767]: I0317 16:27:34.482518 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:27:34 crc kubenswrapper[4767]: E0317 16:27:34.483059 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:27:47 crc kubenswrapper[4767]: I0317 16:27:47.357105 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:27:47 crc kubenswrapper[4767]: E0317 16:27:47.359216 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:27:54 crc kubenswrapper[4767]: I0317 16:27:54.574617 4767 patch_prober.go:28] interesting pod/logging-loki-distributor-9c6b6d984-q9dnr container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 16:27:54 crc kubenswrapper[4767]: I0317 16:27:54.575446 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" podUID="28ac74bb-3f68-478d-8f50-acf61b2e8223" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.53:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 16:27:54 crc kubenswrapper[4767]: I0317 16:27:54.588717 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" podUID="16e4d9a1-285b-4221-8a99-55d515bc3356" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 16:27:54 crc kubenswrapper[4767]: I0317 16:27:54.588841 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" podUID="16e4d9a1-285b-4221-8a99-55d515bc3356" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 16:27:54 crc kubenswrapper[4767]: I0317 16:27:54.593628 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" podUID="0163654c-d57e-4b14-aba0-f76dbaff1114" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 16:27:54 crc kubenswrapper[4767]: E0317 16:27:54.730757 4767 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.219s" Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.161086 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562748-d47vb"] Mar 17 16:28:00 crc kubenswrapper[4767]: E0317 16:28:00.162456 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24066b25-22e0-4da5-add8-ffade9245cb1" containerName="oc" Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.162477 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="24066b25-22e0-4da5-add8-ffade9245cb1" containerName="oc" Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.162841 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="24066b25-22e0-4da5-add8-ffade9245cb1" containerName="oc" Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.164119 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562748-d47vb" Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.167428 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.167550 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.167575 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.175312 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562748-d47vb"] Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.318595 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z29km\" (UniqueName: \"kubernetes.io/projected/52e1bc69-b0e4-43ca-a290-504b2f0fc172-kube-api-access-z29km\") pod \"auto-csr-approver-29562748-d47vb\" (UID: \"52e1bc69-b0e4-43ca-a290-504b2f0fc172\") " pod="openshift-infra/auto-csr-approver-29562748-d47vb" Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.422068 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z29km\" (UniqueName: \"kubernetes.io/projected/52e1bc69-b0e4-43ca-a290-504b2f0fc172-kube-api-access-z29km\") pod \"auto-csr-approver-29562748-d47vb\" (UID: \"52e1bc69-b0e4-43ca-a290-504b2f0fc172\") " pod="openshift-infra/auto-csr-approver-29562748-d47vb" Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.443589 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z29km\" (UniqueName: \"kubernetes.io/projected/52e1bc69-b0e4-43ca-a290-504b2f0fc172-kube-api-access-z29km\") pod \"auto-csr-approver-29562748-d47vb\" (UID: \"52e1bc69-b0e4-43ca-a290-504b2f0fc172\") " pod="openshift-infra/auto-csr-approver-29562748-d47vb" Mar 17 16:28:00 crc kubenswrapper[4767]: I0317 16:28:00.490990 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562748-d47vb" Mar 17 16:28:01 crc kubenswrapper[4767]: I0317 16:28:01.254052 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562748-d47vb"] Mar 17 16:28:02 crc kubenswrapper[4767]: I0317 16:28:02.174677 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562748-d47vb" event={"ID":"52e1bc69-b0e4-43ca-a290-504b2f0fc172","Type":"ContainerStarted","Data":"c463f3953279ab3087dad1885d05b2cb917b071b8ea2ee32899abbe33c80d823"} Mar 17 16:28:02 crc kubenswrapper[4767]: I0317 16:28:02.355043 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:28:02 crc kubenswrapper[4767]: E0317 16:28:02.355889 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:28:03 crc kubenswrapper[4767]: I0317 16:28:03.188574 4767 generic.go:334] "Generic (PLEG): container finished" podID="52e1bc69-b0e4-43ca-a290-504b2f0fc172" containerID="7a5eec1a805d92b9807d1106fb4013ab6a6ae0b1943fd7748470bb2d1ce276fe" exitCode=0 Mar 17 16:28:03 crc kubenswrapper[4767]: I0317 16:28:03.188903 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562748-d47vb" event={"ID":"52e1bc69-b0e4-43ca-a290-504b2f0fc172","Type":"ContainerDied","Data":"7a5eec1a805d92b9807d1106fb4013ab6a6ae0b1943fd7748470bb2d1ce276fe"} Mar 17 16:28:04 crc kubenswrapper[4767]: I0317 16:28:04.850203 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562748-d47vb" Mar 17 16:28:04 crc kubenswrapper[4767]: I0317 16:28:04.911599 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z29km\" (UniqueName: \"kubernetes.io/projected/52e1bc69-b0e4-43ca-a290-504b2f0fc172-kube-api-access-z29km\") pod \"52e1bc69-b0e4-43ca-a290-504b2f0fc172\" (UID: \"52e1bc69-b0e4-43ca-a290-504b2f0fc172\") " Mar 17 16:28:04 crc kubenswrapper[4767]: I0317 16:28:04.918055 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52e1bc69-b0e4-43ca-a290-504b2f0fc172-kube-api-access-z29km" (OuterVolumeSpecName: "kube-api-access-z29km") pod "52e1bc69-b0e4-43ca-a290-504b2f0fc172" (UID: "52e1bc69-b0e4-43ca-a290-504b2f0fc172"). InnerVolumeSpecName "kube-api-access-z29km". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:28:05 crc kubenswrapper[4767]: I0317 16:28:05.013896 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z29km\" (UniqueName: \"kubernetes.io/projected/52e1bc69-b0e4-43ca-a290-504b2f0fc172-kube-api-access-z29km\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:05 crc kubenswrapper[4767]: I0317 16:28:05.218589 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562748-d47vb" event={"ID":"52e1bc69-b0e4-43ca-a290-504b2f0fc172","Type":"ContainerDied","Data":"c463f3953279ab3087dad1885d05b2cb917b071b8ea2ee32899abbe33c80d823"} Mar 17 16:28:05 crc kubenswrapper[4767]: I0317 16:28:05.218642 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c463f3953279ab3087dad1885d05b2cb917b071b8ea2ee32899abbe33c80d823" Mar 17 16:28:05 crc kubenswrapper[4767]: I0317 16:28:05.218708 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562748-d47vb" Mar 17 16:28:05 crc kubenswrapper[4767]: I0317 16:28:05.949072 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562742-77vcr"] Mar 17 16:28:05 crc kubenswrapper[4767]: I0317 16:28:05.961550 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562742-77vcr"] Mar 17 16:28:07 crc kubenswrapper[4767]: I0317 16:28:07.369337 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00f74843-f644-49ff-8799-2a600a1f17b7" path="/var/lib/kubelet/pods/00f74843-f644-49ff-8799-2a600a1f17b7/volumes" Mar 17 16:28:12 crc kubenswrapper[4767]: I0317 16:28:12.500507 4767 generic.go:334] "Generic (PLEG): container finished" podID="bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" containerID="876a640a3ef3261e4ff5c78ee072b8d66a3731829131a4e6fe0e14b067caca64" exitCode=0 Mar 17 16:28:12 crc kubenswrapper[4767]: I0317 16:28:12.500664 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" event={"ID":"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a","Type":"ContainerDied","Data":"876a640a3ef3261e4ff5c78ee072b8d66a3731829131a4e6fe0e14b067caca64"} Mar 17 16:28:13 crc kubenswrapper[4767]: I0317 16:28:13.355321 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:28:13 crc kubenswrapper[4767]: E0317 16:28:13.355855 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.133803 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.259470 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-inventory\") pod \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.260025 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hx8d8\" (UniqueName: \"kubernetes.io/projected/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-kube-api-access-hx8d8\") pod \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.260145 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-combined-ca-bundle\") pod \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.260186 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-0\") pod \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.260241 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-2\") pod \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.260286 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-ssh-key-openstack-edpm-ipam\") pod \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.260310 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-1\") pod \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.260340 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-3\") pod \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.260376 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-0\") pod \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.260482 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-extra-config-0\") pod \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.260651 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-1\") pod \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\" (UID: \"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a\") " Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.268658 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" (UID: "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.270206 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-kube-api-access-hx8d8" (OuterVolumeSpecName: "kube-api-access-hx8d8") pod "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" (UID: "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a"). InnerVolumeSpecName "kube-api-access-hx8d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.304106 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" (UID: "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.314696 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" (UID: "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.316260 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" (UID: "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.327751 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" (UID: "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.329285 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" (UID: "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.339701 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" (UID: "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.344342 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" (UID: "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.348885 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-inventory" (OuterVolumeSpecName: "inventory") pod "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" (UID: "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.359689 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" (UID: "bf5f10cd-4fc8-4bfb-ad93-93d54f17702a"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.364504 4767 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.364551 4767 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.364564 4767 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.364575 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.364584 4767 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.364592 4767 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.364604 4767 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.364616 4767 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.364626 4767 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.364636 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.364647 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hx8d8\" (UniqueName: \"kubernetes.io/projected/bf5f10cd-4fc8-4bfb-ad93-93d54f17702a-kube-api-access-hx8d8\") on node \"crc\" DevicePath \"\"" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.527471 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" event={"ID":"bf5f10cd-4fc8-4bfb-ad93-93d54f17702a","Type":"ContainerDied","Data":"f3cebea0ca8abc3cd021683c0a5439a9cf8feb2cca0f69bad2837e0365ecd30a"} Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.527525 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3cebea0ca8abc3cd021683c0a5439a9cf8feb2cca0f69bad2837e0365ecd30a" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.527562 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-9rvm7" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.710761 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz"] Mar 17 16:28:14 crc kubenswrapper[4767]: E0317 16:28:14.711552 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.711578 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 17 16:28:14 crc kubenswrapper[4767]: E0317 16:28:14.711606 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e1bc69-b0e4-43ca-a290-504b2f0fc172" containerName="oc" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.711613 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e1bc69-b0e4-43ca-a290-504b2f0fc172" containerName="oc" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.711949 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="52e1bc69-b0e4-43ca-a290-504b2f0fc172" containerName="oc" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.711984 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf5f10cd-4fc8-4bfb-ad93-93d54f17702a" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.713160 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.719282 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.719734 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.720208 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.720864 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.720993 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.746318 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz"] Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.775823 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.775996 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.776074 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdh64\" (UniqueName: \"kubernetes.io/projected/a243ab49-3203-461e-8267-b895c6b9318e-kube-api-access-qdh64\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.776224 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.776309 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.776510 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.776783 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.879498 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.879550 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.879611 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.879690 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.879836 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.879896 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.879919 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdh64\" (UniqueName: \"kubernetes.io/projected/a243ab49-3203-461e-8267-b895c6b9318e-kube-api-access-qdh64\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.884192 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.885073 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.885281 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.885463 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.886828 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.887302 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:14 crc kubenswrapper[4767]: I0317 16:28:14.899558 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdh64\" (UniqueName: \"kubernetes.io/projected/a243ab49-3203-461e-8267-b895c6b9318e-kube-api-access-qdh64\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:15 crc kubenswrapper[4767]: I0317 16:28:15.048454 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:28:15 crc kubenswrapper[4767]: I0317 16:28:15.640051 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz"] Mar 17 16:28:16 crc kubenswrapper[4767]: I0317 16:28:16.720367 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" event={"ID":"a243ab49-3203-461e-8267-b895c6b9318e","Type":"ContainerStarted","Data":"760721eb39356b74e3e9291426cd812e0e0263695cf10ab448aedb71dc0f434c"} Mar 17 16:28:16 crc kubenswrapper[4767]: I0317 16:28:16.721047 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" event={"ID":"a243ab49-3203-461e-8267-b895c6b9318e","Type":"ContainerStarted","Data":"487088983bc381594119ef99690ccbed498f8d6e54513acd2df6ac5d256a4a5b"} Mar 17 16:28:16 crc kubenswrapper[4767]: I0317 16:28:16.775356 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" podStartSLOduration=2.339611859 podStartE2EDuration="2.775330138s" podCreationTimestamp="2026-03-17 16:28:14 +0000 UTC" firstStartedPulling="2026-03-17 16:28:15.642675156 +0000 UTC m=+3087.055991203" lastFinishedPulling="2026-03-17 16:28:16.078393435 +0000 UTC m=+3087.491709482" observedRunningTime="2026-03-17 16:28:16.74972804 +0000 UTC m=+3088.163044087" watchObservedRunningTime="2026-03-17 16:28:16.775330138 +0000 UTC m=+3088.188646185" Mar 17 16:28:21 crc kubenswrapper[4767]: I0317 16:28:21.502053 4767 scope.go:117] "RemoveContainer" containerID="b743bd5e5200ce703e9582a3c0c4c8c22ca766ee8284bc2d8cf1ccc9b103407c" Mar 17 16:28:24 crc kubenswrapper[4767]: I0317 16:28:24.356296 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:28:24 crc kubenswrapper[4767]: E0317 16:28:24.357788 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:28:35 crc kubenswrapper[4767]: I0317 16:28:35.356676 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:28:35 crc kubenswrapper[4767]: E0317 16:28:35.357478 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:28:47 crc kubenswrapper[4767]: I0317 16:28:47.355210 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:28:47 crc kubenswrapper[4767]: E0317 16:28:47.356033 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:29:00 crc kubenswrapper[4767]: I0317 16:29:00.354830 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:29:00 crc kubenswrapper[4767]: E0317 16:29:00.355692 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:29:13 crc kubenswrapper[4767]: I0317 16:29:13.355837 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:29:13 crc kubenswrapper[4767]: E0317 16:29:13.357133 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.371356 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m2hv7"] Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.382023 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.390389 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m2hv7"] Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.528707 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-utilities\") pod \"certified-operators-m2hv7\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.528801 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-catalog-content\") pod \"certified-operators-m2hv7\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.528890 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4l6m\" (UniqueName: \"kubernetes.io/projected/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-kube-api-access-w4l6m\") pod \"certified-operators-m2hv7\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.632268 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-utilities\") pod \"certified-operators-m2hv7\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.632352 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-catalog-content\") pod \"certified-operators-m2hv7\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.632404 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4l6m\" (UniqueName: \"kubernetes.io/projected/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-kube-api-access-w4l6m\") pod \"certified-operators-m2hv7\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.632940 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-utilities\") pod \"certified-operators-m2hv7\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.633530 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-catalog-content\") pod \"certified-operators-m2hv7\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.657753 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4l6m\" (UniqueName: \"kubernetes.io/projected/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-kube-api-access-w4l6m\") pod \"certified-operators-m2hv7\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:14 crc kubenswrapper[4767]: I0317 16:29:14.765860 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:15 crc kubenswrapper[4767]: I0317 16:29:15.634458 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m2hv7"] Mar 17 16:29:15 crc kubenswrapper[4767]: W0317 16:29:15.638378 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a5fcc43_fefb_4d7d_aff7_4ec46bdd8100.slice/crio-a1cb17558784b2c6f66b398136e96055c55b57b4b4459e2bd098f06beec8a313 WatchSource:0}: Error finding container a1cb17558784b2c6f66b398136e96055c55b57b4b4459e2bd098f06beec8a313: Status 404 returned error can't find the container with id a1cb17558784b2c6f66b398136e96055c55b57b4b4459e2bd098f06beec8a313 Mar 17 16:29:15 crc kubenswrapper[4767]: I0317 16:29:15.828030 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hv7" event={"ID":"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100","Type":"ContainerStarted","Data":"a1cb17558784b2c6f66b398136e96055c55b57b4b4459e2bd098f06beec8a313"} Mar 17 16:29:16 crc kubenswrapper[4767]: I0317 16:29:16.843030 4767 generic.go:334] "Generic (PLEG): container finished" podID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" containerID="f4f551e4c78f2697bfce6e994eb2f95838ac9d8f02b4bdbeadf37a0c3301c8c5" exitCode=0 Mar 17 16:29:16 crc kubenswrapper[4767]: I0317 16:29:16.843095 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hv7" event={"ID":"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100","Type":"ContainerDied","Data":"f4f551e4c78f2697bfce6e994eb2f95838ac9d8f02b4bdbeadf37a0c3301c8c5"} Mar 17 16:29:16 crc kubenswrapper[4767]: I0317 16:29:16.846155 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 16:29:17 crc kubenswrapper[4767]: I0317 16:29:17.859648 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hv7" event={"ID":"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100","Type":"ContainerStarted","Data":"2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e"} Mar 17 16:29:18 crc kubenswrapper[4767]: I0317 16:29:18.884617 4767 generic.go:334] "Generic (PLEG): container finished" podID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" containerID="2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e" exitCode=0 Mar 17 16:29:18 crc kubenswrapper[4767]: I0317 16:29:18.884882 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hv7" event={"ID":"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100","Type":"ContainerDied","Data":"2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e"} Mar 17 16:29:19 crc kubenswrapper[4767]: I0317 16:29:19.901113 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hv7" event={"ID":"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100","Type":"ContainerStarted","Data":"1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa"} Mar 17 16:29:19 crc kubenswrapper[4767]: I0317 16:29:19.925437 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m2hv7" podStartSLOduration=3.336286618 podStartE2EDuration="5.925383622s" podCreationTimestamp="2026-03-17 16:29:14 +0000 UTC" firstStartedPulling="2026-03-17 16:29:16.845886662 +0000 UTC m=+3148.259202719" lastFinishedPulling="2026-03-17 16:29:19.434983666 +0000 UTC m=+3150.848299723" observedRunningTime="2026-03-17 16:29:19.9222138 +0000 UTC m=+3151.335529877" watchObservedRunningTime="2026-03-17 16:29:19.925383622 +0000 UTC m=+3151.338699669" Mar 17 16:29:24 crc kubenswrapper[4767]: I0317 16:29:24.356311 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:29:24 crc kubenswrapper[4767]: E0317 16:29:24.358313 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:29:24 crc kubenswrapper[4767]: I0317 16:29:24.766241 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:24 crc kubenswrapper[4767]: I0317 16:29:24.766306 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:24 crc kubenswrapper[4767]: I0317 16:29:24.827817 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:25 crc kubenswrapper[4767]: I0317 16:29:25.047310 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:26 crc kubenswrapper[4767]: I0317 16:29:26.352019 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m2hv7"] Mar 17 16:29:27 crc kubenswrapper[4767]: I0317 16:29:27.020346 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m2hv7" podUID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" containerName="registry-server" containerID="cri-o://1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa" gracePeriod=2 Mar 17 16:29:27 crc kubenswrapper[4767]: I0317 16:29:27.629068 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:27 crc kubenswrapper[4767]: I0317 16:29:27.670472 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4l6m\" (UniqueName: \"kubernetes.io/projected/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-kube-api-access-w4l6m\") pod \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " Mar 17 16:29:27 crc kubenswrapper[4767]: I0317 16:29:27.670590 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-catalog-content\") pod \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " Mar 17 16:29:27 crc kubenswrapper[4767]: I0317 16:29:27.670663 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-utilities\") pod \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\" (UID: \"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100\") " Mar 17 16:29:27 crc kubenswrapper[4767]: I0317 16:29:27.672263 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-utilities" (OuterVolumeSpecName: "utilities") pod "0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" (UID: "0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:29:27 crc kubenswrapper[4767]: I0317 16:29:27.679318 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-kube-api-access-w4l6m" (OuterVolumeSpecName: "kube-api-access-w4l6m") pod "0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" (UID: "0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100"). InnerVolumeSpecName "kube-api-access-w4l6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:29:27 crc kubenswrapper[4767]: I0317 16:29:27.943094 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" (UID: "0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:29:27 crc kubenswrapper[4767]: I0317 16:29:27.946132 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:29:27 crc kubenswrapper[4767]: I0317 16:29:27.946156 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4l6m\" (UniqueName: \"kubernetes.io/projected/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-kube-api-access-w4l6m\") on node \"crc\" DevicePath \"\"" Mar 17 16:29:27 crc kubenswrapper[4767]: I0317 16:29:27.946295 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.059220 4767 generic.go:334] "Generic (PLEG): container finished" podID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" containerID="1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa" exitCode=0 Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.059304 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hv7" event={"ID":"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100","Type":"ContainerDied","Data":"1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa"} Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.059355 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hv7" event={"ID":"0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100","Type":"ContainerDied","Data":"a1cb17558784b2c6f66b398136e96055c55b57b4b4459e2bd098f06beec8a313"} Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.059386 4767 scope.go:117] "RemoveContainer" containerID="1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa" Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.061590 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2hv7" Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.097285 4767 scope.go:117] "RemoveContainer" containerID="2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e" Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.129959 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m2hv7"] Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.145413 4767 scope.go:117] "RemoveContainer" containerID="f4f551e4c78f2697bfce6e994eb2f95838ac9d8f02b4bdbeadf37a0c3301c8c5" Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.149328 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m2hv7"] Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.202377 4767 scope.go:117] "RemoveContainer" containerID="1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa" Mar 17 16:29:28 crc kubenswrapper[4767]: E0317 16:29:28.203097 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa\": container with ID starting with 1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa not found: ID does not exist" containerID="1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa" Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.203217 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa"} err="failed to get container status \"1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa\": rpc error: code = NotFound desc = could not find container \"1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa\": container with ID starting with 1293137f41125b024e29d85c77db1ce8857574f6dac92e5f4f027d47dde128aa not found: ID does not exist" Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.203268 4767 scope.go:117] "RemoveContainer" containerID="2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e" Mar 17 16:29:28 crc kubenswrapper[4767]: E0317 16:29:28.204476 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e\": container with ID starting with 2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e not found: ID does not exist" containerID="2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e" Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.204502 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e"} err="failed to get container status \"2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e\": rpc error: code = NotFound desc = could not find container \"2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e\": container with ID starting with 2fc5864c00f4b5af5101eea93aa7c7d5fe715f497548ed2fba7b4a5dca76d67e not found: ID does not exist" Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.204530 4767 scope.go:117] "RemoveContainer" containerID="f4f551e4c78f2697bfce6e994eb2f95838ac9d8f02b4bdbeadf37a0c3301c8c5" Mar 17 16:29:28 crc kubenswrapper[4767]: E0317 16:29:28.204777 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4f551e4c78f2697bfce6e994eb2f95838ac9d8f02b4bdbeadf37a0c3301c8c5\": container with ID starting with f4f551e4c78f2697bfce6e994eb2f95838ac9d8f02b4bdbeadf37a0c3301c8c5 not found: ID does not exist" containerID="f4f551e4c78f2697bfce6e994eb2f95838ac9d8f02b4bdbeadf37a0c3301c8c5" Mar 17 16:29:28 crc kubenswrapper[4767]: I0317 16:29:28.204794 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4f551e4c78f2697bfce6e994eb2f95838ac9d8f02b4bdbeadf37a0c3301c8c5"} err="failed to get container status \"f4f551e4c78f2697bfce6e994eb2f95838ac9d8f02b4bdbeadf37a0c3301c8c5\": rpc error: code = NotFound desc = could not find container \"f4f551e4c78f2697bfce6e994eb2f95838ac9d8f02b4bdbeadf37a0c3301c8c5\": container with ID starting with f4f551e4c78f2697bfce6e994eb2f95838ac9d8f02b4bdbeadf37a0c3301c8c5 not found: ID does not exist" Mar 17 16:29:29 crc kubenswrapper[4767]: I0317 16:29:29.388652 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" path="/var/lib/kubelet/pods/0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100/volumes" Mar 17 16:29:39 crc kubenswrapper[4767]: I0317 16:29:39.364918 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:29:39 crc kubenswrapper[4767]: E0317 16:29:39.365951 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.438652 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kpqjq"] Mar 17 16:29:43 crc kubenswrapper[4767]: E0317 16:29:43.439954 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" containerName="registry-server" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.439971 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" containerName="registry-server" Mar 17 16:29:43 crc kubenswrapper[4767]: E0317 16:29:43.439999 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" containerName="extract-utilities" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.440119 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" containerName="extract-utilities" Mar 17 16:29:43 crc kubenswrapper[4767]: E0317 16:29:43.440140 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" containerName="extract-content" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.440148 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" containerName="extract-content" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.440455 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a5fcc43-fefb-4d7d-aff7-4ec46bdd8100" containerName="registry-server" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.442606 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.457130 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kpqjq"] Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.551766 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkkjp\" (UniqueName: \"kubernetes.io/projected/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-kube-api-access-pkkjp\") pod \"redhat-operators-kpqjq\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.552004 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-catalog-content\") pod \"redhat-operators-kpqjq\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.552472 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-utilities\") pod \"redhat-operators-kpqjq\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.655214 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkkjp\" (UniqueName: \"kubernetes.io/projected/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-kube-api-access-pkkjp\") pod \"redhat-operators-kpqjq\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.655324 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-catalog-content\") pod \"redhat-operators-kpqjq\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.655523 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-utilities\") pod \"redhat-operators-kpqjq\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.656486 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-catalog-content\") pod \"redhat-operators-kpqjq\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.656609 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-utilities\") pod \"redhat-operators-kpqjq\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.680156 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkkjp\" (UniqueName: \"kubernetes.io/projected/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-kube-api-access-pkkjp\") pod \"redhat-operators-kpqjq\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:43 crc kubenswrapper[4767]: I0317 16:29:43.787082 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:44 crc kubenswrapper[4767]: I0317 16:29:44.400785 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kpqjq"] Mar 17 16:29:45 crc kubenswrapper[4767]: I0317 16:29:45.297879 4767 generic.go:334] "Generic (PLEG): container finished" podID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerID="c837e1701836f3f658a88670405fc01bf08d82da7b6e3d2b69936b0f1a138f38" exitCode=0 Mar 17 16:29:45 crc kubenswrapper[4767]: I0317 16:29:45.297974 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpqjq" event={"ID":"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1","Type":"ContainerDied","Data":"c837e1701836f3f658a88670405fc01bf08d82da7b6e3d2b69936b0f1a138f38"} Mar 17 16:29:45 crc kubenswrapper[4767]: I0317 16:29:45.298238 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpqjq" event={"ID":"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1","Type":"ContainerStarted","Data":"bcdf60226a7b5a1599da21ac9d7b9a7aef04658fb082bf1ae30c60148ead8516"} Mar 17 16:29:46 crc kubenswrapper[4767]: I0317 16:29:46.311319 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpqjq" event={"ID":"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1","Type":"ContainerStarted","Data":"881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738"} Mar 17 16:29:50 crc kubenswrapper[4767]: I0317 16:29:50.355927 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:29:50 crc kubenswrapper[4767]: E0317 16:29:50.356720 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:29:50 crc kubenswrapper[4767]: I0317 16:29:50.387391 4767 generic.go:334] "Generic (PLEG): container finished" podID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerID="881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738" exitCode=0 Mar 17 16:29:50 crc kubenswrapper[4767]: I0317 16:29:50.387456 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpqjq" event={"ID":"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1","Type":"ContainerDied","Data":"881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738"} Mar 17 16:29:51 crc kubenswrapper[4767]: I0317 16:29:51.409388 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpqjq" event={"ID":"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1","Type":"ContainerStarted","Data":"5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a"} Mar 17 16:29:51 crc kubenswrapper[4767]: I0317 16:29:51.445908 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kpqjq" podStartSLOduration=2.797204545 podStartE2EDuration="8.445876925s" podCreationTimestamp="2026-03-17 16:29:43 +0000 UTC" firstStartedPulling="2026-03-17 16:29:45.300576851 +0000 UTC m=+3176.713892898" lastFinishedPulling="2026-03-17 16:29:50.949249231 +0000 UTC m=+3182.362565278" observedRunningTime="2026-03-17 16:29:51.439073399 +0000 UTC m=+3182.852389456" watchObservedRunningTime="2026-03-17 16:29:51.445876925 +0000 UTC m=+3182.859192972" Mar 17 16:29:53 crc kubenswrapper[4767]: I0317 16:29:53.787730 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:53 crc kubenswrapper[4767]: I0317 16:29:53.788332 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:29:54 crc kubenswrapper[4767]: I0317 16:29:54.922349 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kpqjq" podUID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerName="registry-server" probeResult="failure" output=< Mar 17 16:29:54 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:29:54 crc kubenswrapper[4767]: > Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.167496 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp"] Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.170852 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.173723 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.174121 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.182641 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562750-nnp6x"] Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.185133 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562750-nnp6x" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.187838 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.187954 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.188092 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.196783 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp"] Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.209734 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562750-nnp6x"] Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.318337 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ba1de78-0522-4d00-bbde-b70b4101b350-config-volume\") pod \"collect-profiles-29562750-758xp\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.318538 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ba1de78-0522-4d00-bbde-b70b4101b350-secret-volume\") pod \"collect-profiles-29562750-758xp\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.318594 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8drf\" (UniqueName: \"kubernetes.io/projected/9ba1de78-0522-4d00-bbde-b70b4101b350-kube-api-access-v8drf\") pod \"collect-profiles-29562750-758xp\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.318828 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xvwc\" (UniqueName: \"kubernetes.io/projected/000288b2-7636-4225-8278-6af17a49cc09-kube-api-access-2xvwc\") pod \"auto-csr-approver-29562750-nnp6x\" (UID: \"000288b2-7636-4225-8278-6af17a49cc09\") " pod="openshift-infra/auto-csr-approver-29562750-nnp6x" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.422371 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ba1de78-0522-4d00-bbde-b70b4101b350-secret-volume\") pod \"collect-profiles-29562750-758xp\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.422499 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8drf\" (UniqueName: \"kubernetes.io/projected/9ba1de78-0522-4d00-bbde-b70b4101b350-kube-api-access-v8drf\") pod \"collect-profiles-29562750-758xp\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.422684 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xvwc\" (UniqueName: \"kubernetes.io/projected/000288b2-7636-4225-8278-6af17a49cc09-kube-api-access-2xvwc\") pod \"auto-csr-approver-29562750-nnp6x\" (UID: \"000288b2-7636-4225-8278-6af17a49cc09\") " pod="openshift-infra/auto-csr-approver-29562750-nnp6x" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.423070 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ba1de78-0522-4d00-bbde-b70b4101b350-config-volume\") pod \"collect-profiles-29562750-758xp\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.425088 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ba1de78-0522-4d00-bbde-b70b4101b350-config-volume\") pod \"collect-profiles-29562750-758xp\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.431734 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ba1de78-0522-4d00-bbde-b70b4101b350-secret-volume\") pod \"collect-profiles-29562750-758xp\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.440015 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xvwc\" (UniqueName: \"kubernetes.io/projected/000288b2-7636-4225-8278-6af17a49cc09-kube-api-access-2xvwc\") pod \"auto-csr-approver-29562750-nnp6x\" (UID: \"000288b2-7636-4225-8278-6af17a49cc09\") " pod="openshift-infra/auto-csr-approver-29562750-nnp6x" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.440633 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8drf\" (UniqueName: \"kubernetes.io/projected/9ba1de78-0522-4d00-bbde-b70b4101b350-kube-api-access-v8drf\") pod \"collect-profiles-29562750-758xp\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.531161 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:00 crc kubenswrapper[4767]: I0317 16:30:00.535768 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562750-nnp6x" Mar 17 16:30:01 crc kubenswrapper[4767]: I0317 16:30:01.091383 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562750-nnp6x"] Mar 17 16:30:01 crc kubenswrapper[4767]: W0317 16:30:01.098645 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ba1de78_0522_4d00_bbde_b70b4101b350.slice/crio-5531ebe3fb5df56e12ff0d3f31f92e6f2a4233f99044880ebf2dc8c986da9862 WatchSource:0}: Error finding container 5531ebe3fb5df56e12ff0d3f31f92e6f2a4233f99044880ebf2dc8c986da9862: Status 404 returned error can't find the container with id 5531ebe3fb5df56e12ff0d3f31f92e6f2a4233f99044880ebf2dc8c986da9862 Mar 17 16:30:01 crc kubenswrapper[4767]: I0317 16:30:01.106503 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp"] Mar 17 16:30:01 crc kubenswrapper[4767]: I0317 16:30:01.547133 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" event={"ID":"9ba1de78-0522-4d00-bbde-b70b4101b350","Type":"ContainerStarted","Data":"6a58c971f23c7a2003e4765beec016510f0c1223f9411dc41b47c605a4d3305e"} Mar 17 16:30:01 crc kubenswrapper[4767]: I0317 16:30:01.547534 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" event={"ID":"9ba1de78-0522-4d00-bbde-b70b4101b350","Type":"ContainerStarted","Data":"5531ebe3fb5df56e12ff0d3f31f92e6f2a4233f99044880ebf2dc8c986da9862"} Mar 17 16:30:01 crc kubenswrapper[4767]: I0317 16:30:01.550111 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562750-nnp6x" event={"ID":"000288b2-7636-4225-8278-6af17a49cc09","Type":"ContainerStarted","Data":"6d968b6185d30e75196c2de27de5e282bc24e8d123ca2ab56cf33d40a1167e1a"} Mar 17 16:30:01 crc kubenswrapper[4767]: I0317 16:30:01.577371 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" podStartSLOduration=1.5773448129999998 podStartE2EDuration="1.577344813s" podCreationTimestamp="2026-03-17 16:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:30:01.565319636 +0000 UTC m=+3192.978635703" watchObservedRunningTime="2026-03-17 16:30:01.577344813 +0000 UTC m=+3192.990660880" Mar 17 16:30:02 crc kubenswrapper[4767]: I0317 16:30:02.566120 4767 generic.go:334] "Generic (PLEG): container finished" podID="9ba1de78-0522-4d00-bbde-b70b4101b350" containerID="6a58c971f23c7a2003e4765beec016510f0c1223f9411dc41b47c605a4d3305e" exitCode=0 Mar 17 16:30:02 crc kubenswrapper[4767]: I0317 16:30:02.566462 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" event={"ID":"9ba1de78-0522-4d00-bbde-b70b4101b350","Type":"ContainerDied","Data":"6a58c971f23c7a2003e4765beec016510f0c1223f9411dc41b47c605a4d3305e"} Mar 17 16:30:03 crc kubenswrapper[4767]: I0317 16:30:03.581128 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562750-nnp6x" event={"ID":"000288b2-7636-4225-8278-6af17a49cc09","Type":"ContainerStarted","Data":"dd982ed329f3980d9a7abefd5211f50bf7815c10156af2e08feaa477b634d16e"} Mar 17 16:30:03 crc kubenswrapper[4767]: I0317 16:30:03.613160 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562750-nnp6x" podStartSLOduration=1.7307996 podStartE2EDuration="3.613137226s" podCreationTimestamp="2026-03-17 16:30:00 +0000 UTC" firstStartedPulling="2026-03-17 16:30:01.099128679 +0000 UTC m=+3192.512444716" lastFinishedPulling="2026-03-17 16:30:02.981466295 +0000 UTC m=+3194.394782342" observedRunningTime="2026-03-17 16:30:03.602100308 +0000 UTC m=+3195.015416375" watchObservedRunningTime="2026-03-17 16:30:03.613137226 +0000 UTC m=+3195.026453263" Mar 17 16:30:03 crc kubenswrapper[4767]: I0317 16:30:03.870054 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:30:03 crc kubenswrapper[4767]: I0317 16:30:03.946702 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.066129 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.145118 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kpqjq"] Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.180915 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ba1de78-0522-4d00-bbde-b70b4101b350-secret-volume\") pod \"9ba1de78-0522-4d00-bbde-b70b4101b350\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.182490 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8drf\" (UniqueName: \"kubernetes.io/projected/9ba1de78-0522-4d00-bbde-b70b4101b350-kube-api-access-v8drf\") pod \"9ba1de78-0522-4d00-bbde-b70b4101b350\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.182724 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ba1de78-0522-4d00-bbde-b70b4101b350-config-volume\") pod \"9ba1de78-0522-4d00-bbde-b70b4101b350\" (UID: \"9ba1de78-0522-4d00-bbde-b70b4101b350\") " Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.183460 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ba1de78-0522-4d00-bbde-b70b4101b350-config-volume" (OuterVolumeSpecName: "config-volume") pod "9ba1de78-0522-4d00-bbde-b70b4101b350" (UID: "9ba1de78-0522-4d00-bbde-b70b4101b350"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.184601 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9ba1de78-0522-4d00-bbde-b70b4101b350-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.189127 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ba1de78-0522-4d00-bbde-b70b4101b350-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9ba1de78-0522-4d00-bbde-b70b4101b350" (UID: "9ba1de78-0522-4d00-bbde-b70b4101b350"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.189331 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ba1de78-0522-4d00-bbde-b70b4101b350-kube-api-access-v8drf" (OuterVolumeSpecName: "kube-api-access-v8drf") pod "9ba1de78-0522-4d00-bbde-b70b4101b350" (UID: "9ba1de78-0522-4d00-bbde-b70b4101b350"). InnerVolumeSpecName "kube-api-access-v8drf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.288270 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8drf\" (UniqueName: \"kubernetes.io/projected/9ba1de78-0522-4d00-bbde-b70b4101b350-kube-api-access-v8drf\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.288310 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9ba1de78-0522-4d00-bbde-b70b4101b350-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.594994 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" event={"ID":"9ba1de78-0522-4d00-bbde-b70b4101b350","Type":"ContainerDied","Data":"5531ebe3fb5df56e12ff0d3f31f92e6f2a4233f99044880ebf2dc8c986da9862"} Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.595309 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5531ebe3fb5df56e12ff0d3f31f92e6f2a4233f99044880ebf2dc8c986da9862" Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.595072 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562750-758xp" Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.602422 4767 generic.go:334] "Generic (PLEG): container finished" podID="000288b2-7636-4225-8278-6af17a49cc09" containerID="dd982ed329f3980d9a7abefd5211f50bf7815c10156af2e08feaa477b634d16e" exitCode=0 Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.602492 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562750-nnp6x" event={"ID":"000288b2-7636-4225-8278-6af17a49cc09","Type":"ContainerDied","Data":"dd982ed329f3980d9a7abefd5211f50bf7815c10156af2e08feaa477b634d16e"} Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.689227 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562705-775td"] Mar 17 16:30:04 crc kubenswrapper[4767]: I0317 16:30:04.710253 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562705-775td"] Mar 17 16:30:05 crc kubenswrapper[4767]: I0317 16:30:05.356864 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:30:05 crc kubenswrapper[4767]: E0317 16:30:05.358467 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:30:05 crc kubenswrapper[4767]: I0317 16:30:05.375805 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd451935-a4a7-41df-aa19-8e7065df2b0c" path="/var/lib/kubelet/pods/dd451935-a4a7-41df-aa19-8e7065df2b0c/volumes" Mar 17 16:30:05 crc kubenswrapper[4767]: I0317 16:30:05.614768 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kpqjq" podUID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerName="registry-server" containerID="cri-o://5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a" gracePeriod=2 Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.158230 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562750-nnp6x" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.169188 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.256161 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-catalog-content\") pod \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.256269 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkkjp\" (UniqueName: \"kubernetes.io/projected/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-kube-api-access-pkkjp\") pod \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.256512 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-utilities\") pod \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\" (UID: \"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1\") " Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.256672 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xvwc\" (UniqueName: \"kubernetes.io/projected/000288b2-7636-4225-8278-6af17a49cc09-kube-api-access-2xvwc\") pod \"000288b2-7636-4225-8278-6af17a49cc09\" (UID: \"000288b2-7636-4225-8278-6af17a49cc09\") " Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.261154 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-utilities" (OuterVolumeSpecName: "utilities") pod "4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" (UID: "4e40305f-c3c6-4b8f-b4de-9876e7eae2d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.277671 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-kube-api-access-pkkjp" (OuterVolumeSpecName: "kube-api-access-pkkjp") pod "4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" (UID: "4e40305f-c3c6-4b8f-b4de-9876e7eae2d1"). InnerVolumeSpecName "kube-api-access-pkkjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.277796 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/000288b2-7636-4225-8278-6af17a49cc09-kube-api-access-2xvwc" (OuterVolumeSpecName: "kube-api-access-2xvwc") pod "000288b2-7636-4225-8278-6af17a49cc09" (UID: "000288b2-7636-4225-8278-6af17a49cc09"). InnerVolumeSpecName "kube-api-access-2xvwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.359972 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.360017 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xvwc\" (UniqueName: \"kubernetes.io/projected/000288b2-7636-4225-8278-6af17a49cc09-kube-api-access-2xvwc\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.360033 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkkjp\" (UniqueName: \"kubernetes.io/projected/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-kube-api-access-pkkjp\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.421109 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" (UID: "4e40305f-c3c6-4b8f-b4de-9876e7eae2d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.462709 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.629881 4767 generic.go:334] "Generic (PLEG): container finished" podID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerID="5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a" exitCode=0 Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.629965 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpqjq" event={"ID":"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1","Type":"ContainerDied","Data":"5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a"} Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.629975 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kpqjq" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.629997 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpqjq" event={"ID":"4e40305f-c3c6-4b8f-b4de-9876e7eae2d1","Type":"ContainerDied","Data":"bcdf60226a7b5a1599da21ac9d7b9a7aef04658fb082bf1ae30c60148ead8516"} Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.630019 4767 scope.go:117] "RemoveContainer" containerID="5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.635015 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562750-nnp6x" event={"ID":"000288b2-7636-4225-8278-6af17a49cc09","Type":"ContainerDied","Data":"6d968b6185d30e75196c2de27de5e282bc24e8d123ca2ab56cf33d40a1167e1a"} Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.635052 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d968b6185d30e75196c2de27de5e282bc24e8d123ca2ab56cf33d40a1167e1a" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.635115 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562750-nnp6x" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.672416 4767 scope.go:117] "RemoveContainer" containerID="881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.691539 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kpqjq"] Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.721921 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kpqjq"] Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.727139 4767 scope.go:117] "RemoveContainer" containerID="c837e1701836f3f658a88670405fc01bf08d82da7b6e3d2b69936b0f1a138f38" Mar 17 16:30:06 crc kubenswrapper[4767]: I0317 16:30:06.738489 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562744-n9hh5"] Mar 17 16:30:07 crc kubenswrapper[4767]: I0317 16:30:07.054712 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562744-n9hh5"] Mar 17 16:30:07 crc kubenswrapper[4767]: I0317 16:30:07.059555 4767 scope.go:117] "RemoveContainer" containerID="5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a" Mar 17 16:30:07 crc kubenswrapper[4767]: E0317 16:30:07.061010 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a\": container with ID starting with 5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a not found: ID does not exist" containerID="5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a" Mar 17 16:30:07 crc kubenswrapper[4767]: I0317 16:30:07.061070 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a"} err="failed to get container status \"5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a\": rpc error: code = NotFound desc = could not find container \"5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a\": container with ID starting with 5055bdebaad2c28ece4690c5214521f230f2ff6f1b5f6ffdecf9964fc44e182a not found: ID does not exist" Mar 17 16:30:07 crc kubenswrapper[4767]: I0317 16:30:07.061113 4767 scope.go:117] "RemoveContainer" containerID="881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738" Mar 17 16:30:07 crc kubenswrapper[4767]: E0317 16:30:07.063618 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738\": container with ID starting with 881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738 not found: ID does not exist" containerID="881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738" Mar 17 16:30:07 crc kubenswrapper[4767]: I0317 16:30:07.063689 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738"} err="failed to get container status \"881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738\": rpc error: code = NotFound desc = could not find container \"881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738\": container with ID starting with 881b6ef824bc517deef718e547879bd2ebbfbc1c19d4799115bae205688eb738 not found: ID does not exist" Mar 17 16:30:07 crc kubenswrapper[4767]: I0317 16:30:07.063726 4767 scope.go:117] "RemoveContainer" containerID="c837e1701836f3f658a88670405fc01bf08d82da7b6e3d2b69936b0f1a138f38" Mar 17 16:30:07 crc kubenswrapper[4767]: E0317 16:30:07.069308 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c837e1701836f3f658a88670405fc01bf08d82da7b6e3d2b69936b0f1a138f38\": container with ID starting with c837e1701836f3f658a88670405fc01bf08d82da7b6e3d2b69936b0f1a138f38 not found: ID does not exist" containerID="c837e1701836f3f658a88670405fc01bf08d82da7b6e3d2b69936b0f1a138f38" Mar 17 16:30:07 crc kubenswrapper[4767]: I0317 16:30:07.069375 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c837e1701836f3f658a88670405fc01bf08d82da7b6e3d2b69936b0f1a138f38"} err="failed to get container status \"c837e1701836f3f658a88670405fc01bf08d82da7b6e3d2b69936b0f1a138f38\": rpc error: code = NotFound desc = could not find container \"c837e1701836f3f658a88670405fc01bf08d82da7b6e3d2b69936b0f1a138f38\": container with ID starting with c837e1701836f3f658a88670405fc01bf08d82da7b6e3d2b69936b0f1a138f38 not found: ID does not exist" Mar 17 16:30:07 crc kubenswrapper[4767]: I0317 16:30:07.368680 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" path="/var/lib/kubelet/pods/4e40305f-c3c6-4b8f-b4de-9876e7eae2d1/volumes" Mar 17 16:30:07 crc kubenswrapper[4767]: I0317 16:30:07.369675 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52c06b1a-98eb-489e-acc4-e51e0bd5bf0e" path="/var/lib/kubelet/pods/52c06b1a-98eb-489e-acc4-e51e0bd5bf0e/volumes" Mar 17 16:30:16 crc kubenswrapper[4767]: I0317 16:30:16.355421 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:30:16 crc kubenswrapper[4767]: E0317 16:30:16.356275 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:30:21 crc kubenswrapper[4767]: I0317 16:30:21.626999 4767 scope.go:117] "RemoveContainer" containerID="1b4446556f867f7b88071dbfee960e4840070cf0d82d6628ef03d011f3a727bc" Mar 17 16:30:21 crc kubenswrapper[4767]: I0317 16:30:21.666606 4767 scope.go:117] "RemoveContainer" containerID="60ade6192c4d4cff1c4810b7a9364170d213c36e2b1bcd75d7d39b7196c2dff5" Mar 17 16:30:27 crc kubenswrapper[4767]: I0317 16:30:27.355820 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:30:27 crc kubenswrapper[4767]: E0317 16:30:27.356709 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:30:41 crc kubenswrapper[4767]: I0317 16:30:41.355669 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:30:41 crc kubenswrapper[4767]: E0317 16:30:41.356524 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:30:46 crc kubenswrapper[4767]: I0317 16:30:46.364196 4767 generic.go:334] "Generic (PLEG): container finished" podID="a243ab49-3203-461e-8267-b895c6b9318e" containerID="760721eb39356b74e3e9291426cd812e0e0263695cf10ab448aedb71dc0f434c" exitCode=0 Mar 17 16:30:46 crc kubenswrapper[4767]: I0317 16:30:46.364291 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" event={"ID":"a243ab49-3203-461e-8267-b895c6b9318e","Type":"ContainerDied","Data":"760721eb39356b74e3e9291426cd812e0e0263695cf10ab448aedb71dc0f434c"} Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.322571 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.393856 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" event={"ID":"a243ab49-3203-461e-8267-b895c6b9318e","Type":"ContainerDied","Data":"487088983bc381594119ef99690ccbed498f8d6e54513acd2df6ac5d256a4a5b"} Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.393909 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="487088983bc381594119ef99690ccbed498f8d6e54513acd2df6ac5d256a4a5b" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.393951 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mdkfz" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.426031 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-telemetry-combined-ca-bundle\") pod \"a243ab49-3203-461e-8267-b895c6b9318e\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.426110 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdh64\" (UniqueName: \"kubernetes.io/projected/a243ab49-3203-461e-8267-b895c6b9318e-kube-api-access-qdh64\") pod \"a243ab49-3203-461e-8267-b895c6b9318e\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.426205 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-inventory\") pod \"a243ab49-3203-461e-8267-b895c6b9318e\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.426404 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-1\") pod \"a243ab49-3203-461e-8267-b895c6b9318e\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.426501 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-0\") pod \"a243ab49-3203-461e-8267-b895c6b9318e\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.426601 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-2\") pod \"a243ab49-3203-461e-8267-b895c6b9318e\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.426658 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ssh-key-openstack-edpm-ipam\") pod \"a243ab49-3203-461e-8267-b895c6b9318e\" (UID: \"a243ab49-3203-461e-8267-b895c6b9318e\") " Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.440605 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a243ab49-3203-461e-8267-b895c6b9318e" (UID: "a243ab49-3203-461e-8267-b895c6b9318e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.444078 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a243ab49-3203-461e-8267-b895c6b9318e-kube-api-access-qdh64" (OuterVolumeSpecName: "kube-api-access-qdh64") pod "a243ab49-3203-461e-8267-b895c6b9318e" (UID: "a243ab49-3203-461e-8267-b895c6b9318e"). InnerVolumeSpecName "kube-api-access-qdh64". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.479611 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "a243ab49-3203-461e-8267-b895c6b9318e" (UID: "a243ab49-3203-461e-8267-b895c6b9318e"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.481092 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a243ab49-3203-461e-8267-b895c6b9318e" (UID: "a243ab49-3203-461e-8267-b895c6b9318e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.481542 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-inventory" (OuterVolumeSpecName: "inventory") pod "a243ab49-3203-461e-8267-b895c6b9318e" (UID: "a243ab49-3203-461e-8267-b895c6b9318e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.482068 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "a243ab49-3203-461e-8267-b895c6b9318e" (UID: "a243ab49-3203-461e-8267-b895c6b9318e"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.530440 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.530473 4767 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.530483 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdh64\" (UniqueName: \"kubernetes.io/projected/a243ab49-3203-461e-8267-b895c6b9318e-kube-api-access-qdh64\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.530494 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.530515 4767 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.530526 4767 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.545780 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw"] Mar 17 16:30:48 crc kubenswrapper[4767]: E0317 16:30:48.546727 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerName="extract-content" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.546767 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerName="extract-content" Mar 17 16:30:48 crc kubenswrapper[4767]: E0317 16:30:48.546779 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba1de78-0522-4d00-bbde-b70b4101b350" containerName="collect-profiles" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.546788 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba1de78-0522-4d00-bbde-b70b4101b350" containerName="collect-profiles" Mar 17 16:30:48 crc kubenswrapper[4767]: E0317 16:30:48.546833 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a243ab49-3203-461e-8267-b895c6b9318e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.546844 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a243ab49-3203-461e-8267-b895c6b9318e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 17 16:30:48 crc kubenswrapper[4767]: E0317 16:30:48.546858 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerName="extract-utilities" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.546871 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerName="extract-utilities" Mar 17 16:30:48 crc kubenswrapper[4767]: E0317 16:30:48.546897 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerName="registry-server" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.546905 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerName="registry-server" Mar 17 16:30:48 crc kubenswrapper[4767]: E0317 16:30:48.546950 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000288b2-7636-4225-8278-6af17a49cc09" containerName="oc" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.546959 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="000288b2-7636-4225-8278-6af17a49cc09" containerName="oc" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.547385 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ba1de78-0522-4d00-bbde-b70b4101b350" containerName="collect-profiles" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.547417 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a243ab49-3203-461e-8267-b895c6b9318e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.547433 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="000288b2-7636-4225-8278-6af17a49cc09" containerName="oc" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.547464 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e40305f-c3c6-4b8f-b4de-9876e7eae2d1" containerName="registry-server" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.548789 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.555815 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.557364 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "a243ab49-3203-461e-8267-b895c6b9318e" (UID: "a243ab49-3203-461e-8267-b895c6b9318e"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.569716 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw"] Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.633628 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.633757 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h6vr\" (UniqueName: \"kubernetes.io/projected/b35afc71-8c38-4be2-85f2-8415656fd3e8-kube-api-access-6h6vr\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.634008 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.634055 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.634138 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ssh-key-openstack-edpm-ipam\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.634195 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.634738 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.635196 4767 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a243ab49-3203-461e-8267-b895c6b9318e-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.738303 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.738444 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h6vr\" (UniqueName: \"kubernetes.io/projected/b35afc71-8c38-4be2-85f2-8415656fd3e8-kube-api-access-6h6vr\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.739136 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.739753 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.739817 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ssh-key-openstack-edpm-ipam\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.739851 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.739939 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.743275 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ssh-key-openstack-edpm-ipam\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.743275 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.743288 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.743539 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.744290 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.744627 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.756635 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h6vr\" (UniqueName: \"kubernetes.io/projected/b35afc71-8c38-4be2-85f2-8415656fd3e8-kube-api-access-6h6vr\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:48 crc kubenswrapper[4767]: I0317 16:30:48.963942 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:30:49 crc kubenswrapper[4767]: I0317 16:30:49.566704 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw"] Mar 17 16:30:49 crc kubenswrapper[4767]: W0317 16:30:49.570907 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb35afc71_8c38_4be2_85f2_8415656fd3e8.slice/crio-4a9d936e7d690353328f571ce8e262961854067173d9926dd4db6679913c57bb WatchSource:0}: Error finding container 4a9d936e7d690353328f571ce8e262961854067173d9926dd4db6679913c57bb: Status 404 returned error can't find the container with id 4a9d936e7d690353328f571ce8e262961854067173d9926dd4db6679913c57bb Mar 17 16:30:50 crc kubenswrapper[4767]: I0317 16:30:50.421200 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" event={"ID":"b35afc71-8c38-4be2-85f2-8415656fd3e8","Type":"ContainerStarted","Data":"4a9d936e7d690353328f571ce8e262961854067173d9926dd4db6679913c57bb"} Mar 17 16:30:51 crc kubenswrapper[4767]: I0317 16:30:51.445994 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" event={"ID":"b35afc71-8c38-4be2-85f2-8415656fd3e8","Type":"ContainerStarted","Data":"2ef456b6342560f88354a94753ce8314769f3cd561f55df512d52579806684ee"} Mar 17 16:30:51 crc kubenswrapper[4767]: I0317 16:30:51.495493 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" podStartSLOduration=2.9553503660000002 podStartE2EDuration="3.49546177s" podCreationTimestamp="2026-03-17 16:30:48 +0000 UTC" firstStartedPulling="2026-03-17 16:30:49.573400933 +0000 UTC m=+3240.986716970" lastFinishedPulling="2026-03-17 16:30:50.113512307 +0000 UTC m=+3241.526828374" observedRunningTime="2026-03-17 16:30:51.4674843 +0000 UTC m=+3242.880800357" watchObservedRunningTime="2026-03-17 16:30:51.49546177 +0000 UTC m=+3242.908777817" Mar 17 16:30:55 crc kubenswrapper[4767]: I0317 16:30:55.358941 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:30:55 crc kubenswrapper[4767]: E0317 16:30:55.360186 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:31:10 crc kubenswrapper[4767]: I0317 16:31:10.357124 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:31:10 crc kubenswrapper[4767]: E0317 16:31:10.358898 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:31:24 crc kubenswrapper[4767]: I0317 16:31:24.355122 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:31:24 crc kubenswrapper[4767]: E0317 16:31:24.356049 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:31:37 crc kubenswrapper[4767]: I0317 16:31:37.359983 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:31:37 crc kubenswrapper[4767]: E0317 16:31:37.361007 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:31:52 crc kubenswrapper[4767]: I0317 16:31:52.354802 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:31:52 crc kubenswrapper[4767]: E0317 16:31:52.355675 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:32:00 crc kubenswrapper[4767]: I0317 16:32:00.180365 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562752-98tkl"] Mar 17 16:32:00 crc kubenswrapper[4767]: I0317 16:32:00.183753 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562752-98tkl" Mar 17 16:32:00 crc kubenswrapper[4767]: I0317 16:32:00.188082 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:32:00 crc kubenswrapper[4767]: I0317 16:32:00.188264 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:32:00 crc kubenswrapper[4767]: I0317 16:32:00.188763 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:32:00 crc kubenswrapper[4767]: I0317 16:32:00.193646 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562752-98tkl"] Mar 17 16:32:00 crc kubenswrapper[4767]: I0317 16:32:00.314105 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndwhf\" (UniqueName: \"kubernetes.io/projected/fc79c7ba-34fb-4553-a163-3a1bfd0c8738-kube-api-access-ndwhf\") pod \"auto-csr-approver-29562752-98tkl\" (UID: \"fc79c7ba-34fb-4553-a163-3a1bfd0c8738\") " pod="openshift-infra/auto-csr-approver-29562752-98tkl" Mar 17 16:32:00 crc kubenswrapper[4767]: I0317 16:32:00.417437 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndwhf\" (UniqueName: \"kubernetes.io/projected/fc79c7ba-34fb-4553-a163-3a1bfd0c8738-kube-api-access-ndwhf\") pod \"auto-csr-approver-29562752-98tkl\" (UID: \"fc79c7ba-34fb-4553-a163-3a1bfd0c8738\") " pod="openshift-infra/auto-csr-approver-29562752-98tkl" Mar 17 16:32:00 crc kubenswrapper[4767]: I0317 16:32:00.437493 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndwhf\" (UniqueName: \"kubernetes.io/projected/fc79c7ba-34fb-4553-a163-3a1bfd0c8738-kube-api-access-ndwhf\") pod \"auto-csr-approver-29562752-98tkl\" (UID: \"fc79c7ba-34fb-4553-a163-3a1bfd0c8738\") " pod="openshift-infra/auto-csr-approver-29562752-98tkl" Mar 17 16:32:00 crc kubenswrapper[4767]: I0317 16:32:00.510376 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562752-98tkl" Mar 17 16:32:01 crc kubenswrapper[4767]: I0317 16:32:01.192791 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562752-98tkl"] Mar 17 16:32:01 crc kubenswrapper[4767]: W0317 16:32:01.197232 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc79c7ba_34fb_4553_a163_3a1bfd0c8738.slice/crio-6a8a3379f105f6a78d017b1406e7449a37cf686c6868d71948e449f05f3a8060 WatchSource:0}: Error finding container 6a8a3379f105f6a78d017b1406e7449a37cf686c6868d71948e449f05f3a8060: Status 404 returned error can't find the container with id 6a8a3379f105f6a78d017b1406e7449a37cf686c6868d71948e449f05f3a8060 Mar 17 16:32:01 crc kubenswrapper[4767]: I0317 16:32:01.539483 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562752-98tkl" event={"ID":"fc79c7ba-34fb-4553-a163-3a1bfd0c8738","Type":"ContainerStarted","Data":"6a8a3379f105f6a78d017b1406e7449a37cf686c6868d71948e449f05f3a8060"} Mar 17 16:32:02 crc kubenswrapper[4767]: I0317 16:32:02.556427 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562752-98tkl" event={"ID":"fc79c7ba-34fb-4553-a163-3a1bfd0c8738","Type":"ContainerStarted","Data":"cf5925bd611fef08bd333b4cec5746d486a74b8de1d6a9291d54393e8bc8a447"} Mar 17 16:32:02 crc kubenswrapper[4767]: I0317 16:32:02.583200 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562752-98tkl" podStartSLOduration=1.6661467380000001 podStartE2EDuration="2.583140416s" podCreationTimestamp="2026-03-17 16:32:00 +0000 UTC" firstStartedPulling="2026-03-17 16:32:01.202657664 +0000 UTC m=+3312.615973711" lastFinishedPulling="2026-03-17 16:32:02.119651332 +0000 UTC m=+3313.532967389" observedRunningTime="2026-03-17 16:32:02.570931852 +0000 UTC m=+3313.984247919" watchObservedRunningTime="2026-03-17 16:32:02.583140416 +0000 UTC m=+3313.996456453" Mar 17 16:32:03 crc kubenswrapper[4767]: I0317 16:32:03.571683 4767 generic.go:334] "Generic (PLEG): container finished" podID="fc79c7ba-34fb-4553-a163-3a1bfd0c8738" containerID="cf5925bd611fef08bd333b4cec5746d486a74b8de1d6a9291d54393e8bc8a447" exitCode=0 Mar 17 16:32:03 crc kubenswrapper[4767]: I0317 16:32:03.571762 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562752-98tkl" event={"ID":"fc79c7ba-34fb-4553-a163-3a1bfd0c8738","Type":"ContainerDied","Data":"cf5925bd611fef08bd333b4cec5746d486a74b8de1d6a9291d54393e8bc8a447"} Mar 17 16:32:05 crc kubenswrapper[4767]: I0317 16:32:05.233152 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562752-98tkl" Mar 17 16:32:05 crc kubenswrapper[4767]: I0317 16:32:05.368658 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndwhf\" (UniqueName: \"kubernetes.io/projected/fc79c7ba-34fb-4553-a163-3a1bfd0c8738-kube-api-access-ndwhf\") pod \"fc79c7ba-34fb-4553-a163-3a1bfd0c8738\" (UID: \"fc79c7ba-34fb-4553-a163-3a1bfd0c8738\") " Mar 17 16:32:05 crc kubenswrapper[4767]: I0317 16:32:05.375359 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc79c7ba-34fb-4553-a163-3a1bfd0c8738-kube-api-access-ndwhf" (OuterVolumeSpecName: "kube-api-access-ndwhf") pod "fc79c7ba-34fb-4553-a163-3a1bfd0c8738" (UID: "fc79c7ba-34fb-4553-a163-3a1bfd0c8738"). InnerVolumeSpecName "kube-api-access-ndwhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:32:05 crc kubenswrapper[4767]: I0317 16:32:05.474904 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndwhf\" (UniqueName: \"kubernetes.io/projected/fc79c7ba-34fb-4553-a163-3a1bfd0c8738-kube-api-access-ndwhf\") on node \"crc\" DevicePath \"\"" Mar 17 16:32:05 crc kubenswrapper[4767]: I0317 16:32:05.596744 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562752-98tkl" event={"ID":"fc79c7ba-34fb-4553-a163-3a1bfd0c8738","Type":"ContainerDied","Data":"6a8a3379f105f6a78d017b1406e7449a37cf686c6868d71948e449f05f3a8060"} Mar 17 16:32:05 crc kubenswrapper[4767]: I0317 16:32:05.596825 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562752-98tkl" Mar 17 16:32:05 crc kubenswrapper[4767]: I0317 16:32:05.596847 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a8a3379f105f6a78d017b1406e7449a37cf686c6868d71948e449f05f3a8060" Mar 17 16:32:05 crc kubenswrapper[4767]: I0317 16:32:05.657038 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562746-d8mln"] Mar 17 16:32:05 crc kubenswrapper[4767]: I0317 16:32:05.670465 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562746-d8mln"] Mar 17 16:32:07 crc kubenswrapper[4767]: I0317 16:32:07.354357 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:32:07 crc kubenswrapper[4767]: E0317 16:32:07.355430 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:32:07 crc kubenswrapper[4767]: I0317 16:32:07.370056 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24066b25-22e0-4da5-add8-ffade9245cb1" path="/var/lib/kubelet/pods/24066b25-22e0-4da5-add8-ffade9245cb1/volumes" Mar 17 16:32:22 crc kubenswrapper[4767]: I0317 16:32:22.033451 4767 scope.go:117] "RemoveContainer" containerID="1d62129ba6a2096431bfd0c6413558229b0461aec5c0b9cb4a753612823c4433" Mar 17 16:32:22 crc kubenswrapper[4767]: I0317 16:32:22.355198 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:32:22 crc kubenswrapper[4767]: E0317 16:32:22.355562 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:32:33 crc kubenswrapper[4767]: I0317 16:32:33.355185 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:32:33 crc kubenswrapper[4767]: E0317 16:32:33.356113 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:32:46 crc kubenswrapper[4767]: I0317 16:32:46.355617 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:32:47 crc kubenswrapper[4767]: I0317 16:32:47.367824 4767 generic.go:334] "Generic (PLEG): container finished" podID="b35afc71-8c38-4be2-85f2-8415656fd3e8" containerID="2ef456b6342560f88354a94753ce8314769f3cd561f55df512d52579806684ee" exitCode=0 Mar 17 16:32:47 crc kubenswrapper[4767]: I0317 16:32:47.377029 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" event={"ID":"b35afc71-8c38-4be2-85f2-8415656fd3e8","Type":"ContainerDied","Data":"2ef456b6342560f88354a94753ce8314769f3cd561f55df512d52579806684ee"} Mar 17 16:32:47 crc kubenswrapper[4767]: I0317 16:32:47.377415 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"a2d2cded09e1efc4abc6d103ad883c0499602c9a3b45371df763f0966d84f955"} Mar 17 16:32:48 crc kubenswrapper[4767]: I0317 16:32:48.988016 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.142923 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h6vr\" (UniqueName: \"kubernetes.io/projected/b35afc71-8c38-4be2-85f2-8415656fd3e8-kube-api-access-6h6vr\") pod \"b35afc71-8c38-4be2-85f2-8415656fd3e8\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.142985 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-2\") pod \"b35afc71-8c38-4be2-85f2-8415656fd3e8\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.143066 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ssh-key-openstack-edpm-ipam\") pod \"b35afc71-8c38-4be2-85f2-8415656fd3e8\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.143110 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-0\") pod \"b35afc71-8c38-4be2-85f2-8415656fd3e8\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.143163 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-1\") pod \"b35afc71-8c38-4be2-85f2-8415656fd3e8\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.143375 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-inventory\") pod \"b35afc71-8c38-4be2-85f2-8415656fd3e8\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.143462 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-telemetry-power-monitoring-combined-ca-bundle\") pod \"b35afc71-8c38-4be2-85f2-8415656fd3e8\" (UID: \"b35afc71-8c38-4be2-85f2-8415656fd3e8\") " Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.168029 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "b35afc71-8c38-4be2-85f2-8415656fd3e8" (UID: "b35afc71-8c38-4be2-85f2-8415656fd3e8"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.170681 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b35afc71-8c38-4be2-85f2-8415656fd3e8-kube-api-access-6h6vr" (OuterVolumeSpecName: "kube-api-access-6h6vr") pod "b35afc71-8c38-4be2-85f2-8415656fd3e8" (UID: "b35afc71-8c38-4be2-85f2-8415656fd3e8"). InnerVolumeSpecName "kube-api-access-6h6vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.209467 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "b35afc71-8c38-4be2-85f2-8415656fd3e8" (UID: "b35afc71-8c38-4be2-85f2-8415656fd3e8"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.241507 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "b35afc71-8c38-4be2-85f2-8415656fd3e8" (UID: "b35afc71-8c38-4be2-85f2-8415656fd3e8"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.242378 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-inventory" (OuterVolumeSpecName: "inventory") pod "b35afc71-8c38-4be2-85f2-8415656fd3e8" (UID: "b35afc71-8c38-4be2-85f2-8415656fd3e8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.248195 4767 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.248263 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h6vr\" (UniqueName: \"kubernetes.io/projected/b35afc71-8c38-4be2-85f2-8415656fd3e8-kube-api-access-6h6vr\") on node \"crc\" DevicePath \"\"" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.248278 4767 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.248298 4767 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.248309 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.285340 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b35afc71-8c38-4be2-85f2-8415656fd3e8" (UID: "b35afc71-8c38-4be2-85f2-8415656fd3e8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.291159 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "b35afc71-8c38-4be2-85f2-8415656fd3e8" (UID: "b35afc71-8c38-4be2-85f2-8415656fd3e8"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.351184 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.351230 4767 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/b35afc71-8c38-4be2-85f2-8415656fd3e8-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.398950 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" event={"ID":"b35afc71-8c38-4be2-85f2-8415656fd3e8","Type":"ContainerDied","Data":"4a9d936e7d690353328f571ce8e262961854067173d9926dd4db6679913c57bb"} Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.399328 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a9d936e7d690353328f571ce8e262961854067173d9926dd4db6679913c57bb" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.399029 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wxftw" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.542283 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447"] Mar 17 16:32:49 crc kubenswrapper[4767]: E0317 16:32:49.543118 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b35afc71-8c38-4be2-85f2-8415656fd3e8" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.543143 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b35afc71-8c38-4be2-85f2-8415656fd3e8" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Mar 17 16:32:49 crc kubenswrapper[4767]: E0317 16:32:49.543212 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc79c7ba-34fb-4553-a163-3a1bfd0c8738" containerName="oc" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.543222 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc79c7ba-34fb-4553-a163-3a1bfd0c8738" containerName="oc" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.543485 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="b35afc71-8c38-4be2-85f2-8415656fd3e8" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.543515 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc79c7ba-34fb-4553-a163-3a1bfd0c8738" containerName="oc" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.546073 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.552880 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-489fx" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.552955 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.552975 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.552888 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.552904 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.556055 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447"] Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.664210 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.665999 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.666352 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-ssh-key-openstack-edpm-ipam\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.666695 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.666947 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld85s\" (UniqueName: \"kubernetes.io/projected/962daa90-0720-4162-8187-e93842cc2d76-kube-api-access-ld85s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.768629 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.768710 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-ssh-key-openstack-edpm-ipam\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.768851 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.768962 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld85s\" (UniqueName: \"kubernetes.io/projected/962daa90-0720-4162-8187-e93842cc2d76-kube-api-access-ld85s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.769017 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.775489 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-ssh-key-openstack-edpm-ipam\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.775962 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.776067 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.776706 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.788430 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld85s\" (UniqueName: \"kubernetes.io/projected/962daa90-0720-4162-8187-e93842cc2d76-kube-api-access-ld85s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-kl447\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:49 crc kubenswrapper[4767]: I0317 16:32:49.889486 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:32:50 crc kubenswrapper[4767]: I0317 16:32:50.494339 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447"] Mar 17 16:32:50 crc kubenswrapper[4767]: W0317 16:32:50.495457 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod962daa90_0720_4162_8187_e93842cc2d76.slice/crio-818772cee7c7db9fe3df281564e0790392760f9db49858f0474fa7f5a9543f94 WatchSource:0}: Error finding container 818772cee7c7db9fe3df281564e0790392760f9db49858f0474fa7f5a9543f94: Status 404 returned error can't find the container with id 818772cee7c7db9fe3df281564e0790392760f9db49858f0474fa7f5a9543f94 Mar 17 16:32:50 crc kubenswrapper[4767]: I0317 16:32:50.909965 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 16:32:51 crc kubenswrapper[4767]: I0317 16:32:51.437291 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" event={"ID":"962daa90-0720-4162-8187-e93842cc2d76","Type":"ContainerStarted","Data":"607d3e1e3d97a0c74800854faf8c187096317e2e5cb8993b4a9eb449a9350a57"} Mar 17 16:32:51 crc kubenswrapper[4767]: I0317 16:32:51.437662 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" event={"ID":"962daa90-0720-4162-8187-e93842cc2d76","Type":"ContainerStarted","Data":"818772cee7c7db9fe3df281564e0790392760f9db49858f0474fa7f5a9543f94"} Mar 17 16:32:51 crc kubenswrapper[4767]: I0317 16:32:51.470377 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" podStartSLOduration=2.062076435 podStartE2EDuration="2.470343018s" podCreationTimestamp="2026-03-17 16:32:49 +0000 UTC" firstStartedPulling="2026-03-17 16:32:50.498574253 +0000 UTC m=+3361.911890300" lastFinishedPulling="2026-03-17 16:32:50.906840836 +0000 UTC m=+3362.320156883" observedRunningTime="2026-03-17 16:32:51.461559113 +0000 UTC m=+3362.874875180" watchObservedRunningTime="2026-03-17 16:32:51.470343018 +0000 UTC m=+3362.883659065" Mar 17 16:33:05 crc kubenswrapper[4767]: I0317 16:33:05.863440 4767 generic.go:334] "Generic (PLEG): container finished" podID="962daa90-0720-4162-8187-e93842cc2d76" containerID="607d3e1e3d97a0c74800854faf8c187096317e2e5cb8993b4a9eb449a9350a57" exitCode=0 Mar 17 16:33:05 crc kubenswrapper[4767]: I0317 16:33:05.863630 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" event={"ID":"962daa90-0720-4162-8187-e93842cc2d76","Type":"ContainerDied","Data":"607d3e1e3d97a0c74800854faf8c187096317e2e5cb8993b4a9eb449a9350a57"} Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.368957 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.542684 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-1\") pod \"962daa90-0720-4162-8187-e93842cc2d76\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.542746 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-0\") pod \"962daa90-0720-4162-8187-e93842cc2d76\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.542793 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-ssh-key-openstack-edpm-ipam\") pod \"962daa90-0720-4162-8187-e93842cc2d76\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.542890 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-inventory\") pod \"962daa90-0720-4162-8187-e93842cc2d76\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.542991 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld85s\" (UniqueName: \"kubernetes.io/projected/962daa90-0720-4162-8187-e93842cc2d76-kube-api-access-ld85s\") pod \"962daa90-0720-4162-8187-e93842cc2d76\" (UID: \"962daa90-0720-4162-8187-e93842cc2d76\") " Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.584644 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/962daa90-0720-4162-8187-e93842cc2d76-kube-api-access-ld85s" (OuterVolumeSpecName: "kube-api-access-ld85s") pod "962daa90-0720-4162-8187-e93842cc2d76" (UID: "962daa90-0720-4162-8187-e93842cc2d76"). InnerVolumeSpecName "kube-api-access-ld85s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.610416 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-inventory" (OuterVolumeSpecName: "inventory") pod "962daa90-0720-4162-8187-e93842cc2d76" (UID: "962daa90-0720-4162-8187-e93842cc2d76"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.656311 4767 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.656584 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld85s\" (UniqueName: \"kubernetes.io/projected/962daa90-0720-4162-8187-e93842cc2d76-kube-api-access-ld85s\") on node \"crc\" DevicePath \"\"" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.659490 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "962daa90-0720-4162-8187-e93842cc2d76" (UID: "962daa90-0720-4162-8187-e93842cc2d76"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.664116 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "962daa90-0720-4162-8187-e93842cc2d76" (UID: "962daa90-0720-4162-8187-e93842cc2d76"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.673470 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "962daa90-0720-4162-8187-e93842cc2d76" (UID: "962daa90-0720-4162-8187-e93842cc2d76"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.759115 4767 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.759153 4767 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.759182 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/962daa90-0720-4162-8187-e93842cc2d76-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.892377 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" event={"ID":"962daa90-0720-4162-8187-e93842cc2d76","Type":"ContainerDied","Data":"818772cee7c7db9fe3df281564e0790392760f9db49858f0474fa7f5a9543f94"} Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.892456 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="818772cee7c7db9fe3df281564e0790392760f9db49858f0474fa7f5a9543f94" Mar 17 16:33:07 crc kubenswrapper[4767]: I0317 16:33:07.892548 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-kl447" Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.187799 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562754-mtrx5"] Mar 17 16:34:00 crc kubenswrapper[4767]: E0317 16:34:00.189129 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="962daa90-0720-4162-8187-e93842cc2d76" containerName="logging-edpm-deployment-openstack-edpm-ipam" Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.189155 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="962daa90-0720-4162-8187-e93842cc2d76" containerName="logging-edpm-deployment-openstack-edpm-ipam" Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.189586 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="962daa90-0720-4162-8187-e93842cc2d76" containerName="logging-edpm-deployment-openstack-edpm-ipam" Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.191447 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562754-mtrx5" Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.195064 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.195823 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.196240 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.203776 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562754-mtrx5"] Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.219866 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xddxc\" (UniqueName: \"kubernetes.io/projected/87c15745-87fb-4aff-8065-89ad1efe0f6a-kube-api-access-xddxc\") pod \"auto-csr-approver-29562754-mtrx5\" (UID: \"87c15745-87fb-4aff-8065-89ad1efe0f6a\") " pod="openshift-infra/auto-csr-approver-29562754-mtrx5" Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.327325 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xddxc\" (UniqueName: \"kubernetes.io/projected/87c15745-87fb-4aff-8065-89ad1efe0f6a-kube-api-access-xddxc\") pod \"auto-csr-approver-29562754-mtrx5\" (UID: \"87c15745-87fb-4aff-8065-89ad1efe0f6a\") " pod="openshift-infra/auto-csr-approver-29562754-mtrx5" Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.347082 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xddxc\" (UniqueName: \"kubernetes.io/projected/87c15745-87fb-4aff-8065-89ad1efe0f6a-kube-api-access-xddxc\") pod \"auto-csr-approver-29562754-mtrx5\" (UID: \"87c15745-87fb-4aff-8065-89ad1efe0f6a\") " pod="openshift-infra/auto-csr-approver-29562754-mtrx5" Mar 17 16:34:00 crc kubenswrapper[4767]: I0317 16:34:00.528878 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562754-mtrx5" Mar 17 16:34:01 crc kubenswrapper[4767]: I0317 16:34:01.019450 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562754-mtrx5"] Mar 17 16:34:01 crc kubenswrapper[4767]: I0317 16:34:01.687646 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562754-mtrx5" event={"ID":"87c15745-87fb-4aff-8065-89ad1efe0f6a","Type":"ContainerStarted","Data":"9baed3cecc96d2142bb87099cc2e6304976105a455b6c074cd8eb2d4f66d6339"} Mar 17 16:34:02 crc kubenswrapper[4767]: I0317 16:34:02.702714 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562754-mtrx5" event={"ID":"87c15745-87fb-4aff-8065-89ad1efe0f6a","Type":"ContainerStarted","Data":"cbc242ab350f19501fc6801844c592b6682389281cdc667fef45c953f31febc4"} Mar 17 16:34:02 crc kubenswrapper[4767]: I0317 16:34:02.728915 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562754-mtrx5" podStartSLOduration=1.5373646490000001 podStartE2EDuration="2.728391288s" podCreationTimestamp="2026-03-17 16:34:00 +0000 UTC" firstStartedPulling="2026-03-17 16:34:01.012206898 +0000 UTC m=+3432.425522955" lastFinishedPulling="2026-03-17 16:34:02.203233547 +0000 UTC m=+3433.616549594" observedRunningTime="2026-03-17 16:34:02.718547693 +0000 UTC m=+3434.131863750" watchObservedRunningTime="2026-03-17 16:34:02.728391288 +0000 UTC m=+3434.141707335" Mar 17 16:34:03 crc kubenswrapper[4767]: I0317 16:34:03.717781 4767 generic.go:334] "Generic (PLEG): container finished" podID="87c15745-87fb-4aff-8065-89ad1efe0f6a" containerID="cbc242ab350f19501fc6801844c592b6682389281cdc667fef45c953f31febc4" exitCode=0 Mar 17 16:34:03 crc kubenswrapper[4767]: I0317 16:34:03.717895 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562754-mtrx5" event={"ID":"87c15745-87fb-4aff-8065-89ad1efe0f6a","Type":"ContainerDied","Data":"cbc242ab350f19501fc6801844c592b6682389281cdc667fef45c953f31febc4"} Mar 17 16:34:05 crc kubenswrapper[4767]: I0317 16:34:05.316229 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562754-mtrx5" Mar 17 16:34:05 crc kubenswrapper[4767]: I0317 16:34:05.697879 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xddxc\" (UniqueName: \"kubernetes.io/projected/87c15745-87fb-4aff-8065-89ad1efe0f6a-kube-api-access-xddxc\") pod \"87c15745-87fb-4aff-8065-89ad1efe0f6a\" (UID: \"87c15745-87fb-4aff-8065-89ad1efe0f6a\") " Mar 17 16:34:05 crc kubenswrapper[4767]: I0317 16:34:05.716716 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87c15745-87fb-4aff-8065-89ad1efe0f6a-kube-api-access-xddxc" (OuterVolumeSpecName: "kube-api-access-xddxc") pod "87c15745-87fb-4aff-8065-89ad1efe0f6a" (UID: "87c15745-87fb-4aff-8065-89ad1efe0f6a"). InnerVolumeSpecName "kube-api-access-xddxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:34:05 crc kubenswrapper[4767]: I0317 16:34:05.762928 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562754-mtrx5" event={"ID":"87c15745-87fb-4aff-8065-89ad1efe0f6a","Type":"ContainerDied","Data":"9baed3cecc96d2142bb87099cc2e6304976105a455b6c074cd8eb2d4f66d6339"} Mar 17 16:34:05 crc kubenswrapper[4767]: I0317 16:34:05.763295 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9baed3cecc96d2142bb87099cc2e6304976105a455b6c074cd8eb2d4f66d6339" Mar 17 16:34:05 crc kubenswrapper[4767]: I0317 16:34:05.763385 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562754-mtrx5" Mar 17 16:34:05 crc kubenswrapper[4767]: I0317 16:34:05.814629 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xddxc\" (UniqueName: \"kubernetes.io/projected/87c15745-87fb-4aff-8065-89ad1efe0f6a-kube-api-access-xddxc\") on node \"crc\" DevicePath \"\"" Mar 17 16:34:05 crc kubenswrapper[4767]: I0317 16:34:05.836052 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562748-d47vb"] Mar 17 16:34:05 crc kubenswrapper[4767]: I0317 16:34:05.860448 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562748-d47vb"] Mar 17 16:34:07 crc kubenswrapper[4767]: I0317 16:34:07.369389 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52e1bc69-b0e4-43ca-a290-504b2f0fc172" path="/var/lib/kubelet/pods/52e1bc69-b0e4-43ca-a290-504b2f0fc172/volumes" Mar 17 16:34:22 crc kubenswrapper[4767]: I0317 16:34:22.174843 4767 scope.go:117] "RemoveContainer" containerID="7a5eec1a805d92b9807d1106fb4013ab6a6ae0b1943fd7748470bb2d1ce276fe" Mar 17 16:34:44 crc kubenswrapper[4767]: I0317 16:34:44.987710 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5tr6k"] Mar 17 16:34:44 crc kubenswrapper[4767]: E0317 16:34:44.989345 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c15745-87fb-4aff-8065-89ad1efe0f6a" containerName="oc" Mar 17 16:34:44 crc kubenswrapper[4767]: I0317 16:34:44.989760 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c15745-87fb-4aff-8065-89ad1efe0f6a" containerName="oc" Mar 17 16:34:44 crc kubenswrapper[4767]: I0317 16:34:44.990100 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c15745-87fb-4aff-8065-89ad1efe0f6a" containerName="oc" Mar 17 16:34:44 crc kubenswrapper[4767]: I0317 16:34:44.993213 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:45 crc kubenswrapper[4767]: I0317 16:34:45.011005 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5tr6k"] Mar 17 16:34:45 crc kubenswrapper[4767]: I0317 16:34:45.130250 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1f08cb-7205-4537-acca-91edee17aef9-utilities\") pod \"community-operators-5tr6k\" (UID: \"cc1f08cb-7205-4537-acca-91edee17aef9\") " pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:45 crc kubenswrapper[4767]: I0317 16:34:45.130993 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1f08cb-7205-4537-acca-91edee17aef9-catalog-content\") pod \"community-operators-5tr6k\" (UID: \"cc1f08cb-7205-4537-acca-91edee17aef9\") " pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:45 crc kubenswrapper[4767]: I0317 16:34:45.131201 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92ndh\" (UniqueName: \"kubernetes.io/projected/cc1f08cb-7205-4537-acca-91edee17aef9-kube-api-access-92ndh\") pod \"community-operators-5tr6k\" (UID: \"cc1f08cb-7205-4537-acca-91edee17aef9\") " pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:45 crc kubenswrapper[4767]: I0317 16:34:45.233645 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1f08cb-7205-4537-acca-91edee17aef9-utilities\") pod \"community-operators-5tr6k\" (UID: \"cc1f08cb-7205-4537-acca-91edee17aef9\") " pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:45 crc kubenswrapper[4767]: I0317 16:34:45.233855 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1f08cb-7205-4537-acca-91edee17aef9-catalog-content\") pod \"community-operators-5tr6k\" (UID: \"cc1f08cb-7205-4537-acca-91edee17aef9\") " pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:45 crc kubenswrapper[4767]: I0317 16:34:45.233959 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92ndh\" (UniqueName: \"kubernetes.io/projected/cc1f08cb-7205-4537-acca-91edee17aef9-kube-api-access-92ndh\") pod \"community-operators-5tr6k\" (UID: \"cc1f08cb-7205-4537-acca-91edee17aef9\") " pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:45 crc kubenswrapper[4767]: I0317 16:34:45.234647 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1f08cb-7205-4537-acca-91edee17aef9-catalog-content\") pod \"community-operators-5tr6k\" (UID: \"cc1f08cb-7205-4537-acca-91edee17aef9\") " pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:45 crc kubenswrapper[4767]: I0317 16:34:45.234986 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1f08cb-7205-4537-acca-91edee17aef9-utilities\") pod \"community-operators-5tr6k\" (UID: \"cc1f08cb-7205-4537-acca-91edee17aef9\") " pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:45 crc kubenswrapper[4767]: I0317 16:34:45.259012 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92ndh\" (UniqueName: \"kubernetes.io/projected/cc1f08cb-7205-4537-acca-91edee17aef9-kube-api-access-92ndh\") pod \"community-operators-5tr6k\" (UID: \"cc1f08cb-7205-4537-acca-91edee17aef9\") " pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:45 crc kubenswrapper[4767]: I0317 16:34:45.324163 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:46 crc kubenswrapper[4767]: I0317 16:34:46.177137 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5tr6k"] Mar 17 16:34:46 crc kubenswrapper[4767]: I0317 16:34:46.490295 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5tr6k" event={"ID":"cc1f08cb-7205-4537-acca-91edee17aef9","Type":"ContainerStarted","Data":"1b9350b57a1f8352ba86ae67cd4537fb82b71547c01b9a039c4de85ced61fb17"} Mar 17 16:34:46 crc kubenswrapper[4767]: I0317 16:34:46.490681 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5tr6k" event={"ID":"cc1f08cb-7205-4537-acca-91edee17aef9","Type":"ContainerStarted","Data":"b63c656c46b493c2ea6522dea08468646fd5838571e2bc1d57d270b172274877"} Mar 17 16:34:47 crc kubenswrapper[4767]: I0317 16:34:47.505239 4767 generic.go:334] "Generic (PLEG): container finished" podID="cc1f08cb-7205-4537-acca-91edee17aef9" containerID="1b9350b57a1f8352ba86ae67cd4537fb82b71547c01b9a039c4de85ced61fb17" exitCode=0 Mar 17 16:34:47 crc kubenswrapper[4767]: I0317 16:34:47.505305 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5tr6k" event={"ID":"cc1f08cb-7205-4537-acca-91edee17aef9","Type":"ContainerDied","Data":"1b9350b57a1f8352ba86ae67cd4537fb82b71547c01b9a039c4de85ced61fb17"} Mar 17 16:34:47 crc kubenswrapper[4767]: I0317 16:34:47.508142 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 16:34:53 crc kubenswrapper[4767]: I0317 16:34:53.595481 4767 generic.go:334] "Generic (PLEG): container finished" podID="cc1f08cb-7205-4537-acca-91edee17aef9" containerID="6550f507cd29f4afe253acf979026b390d24d2f0dca89343a26d50f34e5bf7ae" exitCode=0 Mar 17 16:34:53 crc kubenswrapper[4767]: I0317 16:34:53.595974 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5tr6k" event={"ID":"cc1f08cb-7205-4537-acca-91edee17aef9","Type":"ContainerDied","Data":"6550f507cd29f4afe253acf979026b390d24d2f0dca89343a26d50f34e5bf7ae"} Mar 17 16:34:54 crc kubenswrapper[4767]: I0317 16:34:54.611911 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5tr6k" event={"ID":"cc1f08cb-7205-4537-acca-91edee17aef9","Type":"ContainerStarted","Data":"8cebe2fcdf1b70b6be9ba74a0ab06fcb34e2453ea0e94998168f1150c9408680"} Mar 17 16:34:54 crc kubenswrapper[4767]: I0317 16:34:54.654108 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5tr6k" podStartSLOduration=4.12524557 podStartE2EDuration="10.654079016s" podCreationTimestamp="2026-03-17 16:34:44 +0000 UTC" firstStartedPulling="2026-03-17 16:34:47.507873477 +0000 UTC m=+3478.921189524" lastFinishedPulling="2026-03-17 16:34:54.036706923 +0000 UTC m=+3485.450022970" observedRunningTime="2026-03-17 16:34:54.638864275 +0000 UTC m=+3486.052180342" watchObservedRunningTime="2026-03-17 16:34:54.654079016 +0000 UTC m=+3486.067395063" Mar 17 16:34:55 crc kubenswrapper[4767]: I0317 16:34:55.325146 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:55 crc kubenswrapper[4767]: I0317 16:34:55.325232 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:34:56 crc kubenswrapper[4767]: I0317 16:34:56.377457 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-5tr6k" podUID="cc1f08cb-7205-4537-acca-91edee17aef9" containerName="registry-server" probeResult="failure" output=< Mar 17 16:34:56 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:34:56 crc kubenswrapper[4767]: > Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.176821 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l2jhp"] Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.180620 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.221758 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2jhp"] Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.244849 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-utilities\") pod \"redhat-marketplace-l2jhp\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.245069 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk2p4\" (UniqueName: \"kubernetes.io/projected/5dbf9e0d-191f-4b28-be16-19a915eb94e4-kube-api-access-dk2p4\") pod \"redhat-marketplace-l2jhp\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.245096 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-catalog-content\") pod \"redhat-marketplace-l2jhp\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.347430 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk2p4\" (UniqueName: \"kubernetes.io/projected/5dbf9e0d-191f-4b28-be16-19a915eb94e4-kube-api-access-dk2p4\") pod \"redhat-marketplace-l2jhp\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.347819 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-catalog-content\") pod \"redhat-marketplace-l2jhp\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.348030 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-utilities\") pod \"redhat-marketplace-l2jhp\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.348427 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-catalog-content\") pod \"redhat-marketplace-l2jhp\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.348712 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-utilities\") pod \"redhat-marketplace-l2jhp\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.377721 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk2p4\" (UniqueName: \"kubernetes.io/projected/5dbf9e0d-191f-4b28-be16-19a915eb94e4-kube-api-access-dk2p4\") pod \"redhat-marketplace-l2jhp\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:03 crc kubenswrapper[4767]: I0317 16:35:03.521130 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:04 crc kubenswrapper[4767]: I0317 16:35:04.121192 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2jhp"] Mar 17 16:35:04 crc kubenswrapper[4767]: I0317 16:35:04.166142 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:35:04 crc kubenswrapper[4767]: I0317 16:35:04.166228 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:35:04 crc kubenswrapper[4767]: I0317 16:35:04.760714 4767 generic.go:334] "Generic (PLEG): container finished" podID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" containerID="5b9c96e0bcc9e06dcc47ceec01e9d59647c2d284829492aebd7cbd7966234e5a" exitCode=0 Mar 17 16:35:04 crc kubenswrapper[4767]: I0317 16:35:04.760809 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2jhp" event={"ID":"5dbf9e0d-191f-4b28-be16-19a915eb94e4","Type":"ContainerDied","Data":"5b9c96e0bcc9e06dcc47ceec01e9d59647c2d284829492aebd7cbd7966234e5a"} Mar 17 16:35:04 crc kubenswrapper[4767]: I0317 16:35:04.761065 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2jhp" event={"ID":"5dbf9e0d-191f-4b28-be16-19a915eb94e4","Type":"ContainerStarted","Data":"ea9d5723cf82270bb2632b50406a82e3f21f7df76330b4f90793967691a11c6c"} Mar 17 16:35:05 crc kubenswrapper[4767]: I0317 16:35:05.380714 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:35:05 crc kubenswrapper[4767]: I0317 16:35:05.446739 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5tr6k" Mar 17 16:35:06 crc kubenswrapper[4767]: I0317 16:35:06.789959 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2jhp" event={"ID":"5dbf9e0d-191f-4b28-be16-19a915eb94e4","Type":"ContainerStarted","Data":"ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081"} Mar 17 16:35:07 crc kubenswrapper[4767]: I0317 16:35:07.386472 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5tr6k"] Mar 17 16:35:07 crc kubenswrapper[4767]: I0317 16:35:07.756116 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dsjwf"] Mar 17 16:35:07 crc kubenswrapper[4767]: I0317 16:35:07.756964 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dsjwf" podUID="7a3d6a4a-3683-4036-8171-b6488fc6b217" containerName="registry-server" containerID="cri-o://31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd" gracePeriod=2 Mar 17 16:35:07 crc kubenswrapper[4767]: I0317 16:35:07.812890 4767 generic.go:334] "Generic (PLEG): container finished" podID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" containerID="ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081" exitCode=0 Mar 17 16:35:07 crc kubenswrapper[4767]: I0317 16:35:07.812966 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2jhp" event={"ID":"5dbf9e0d-191f-4b28-be16-19a915eb94e4","Type":"ContainerDied","Data":"ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081"} Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.423636 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dsjwf" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.616906 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-utilities\") pod \"7a3d6a4a-3683-4036-8171-b6488fc6b217\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.617423 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-catalog-content\") pod \"7a3d6a4a-3683-4036-8171-b6488fc6b217\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.617596 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk9kx\" (UniqueName: \"kubernetes.io/projected/7a3d6a4a-3683-4036-8171-b6488fc6b217-kube-api-access-vk9kx\") pod \"7a3d6a4a-3683-4036-8171-b6488fc6b217\" (UID: \"7a3d6a4a-3683-4036-8171-b6488fc6b217\") " Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.618564 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-utilities" (OuterVolumeSpecName: "utilities") pod "7a3d6a4a-3683-4036-8171-b6488fc6b217" (UID: "7a3d6a4a-3683-4036-8171-b6488fc6b217"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.620270 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.625497 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a3d6a4a-3683-4036-8171-b6488fc6b217-kube-api-access-vk9kx" (OuterVolumeSpecName: "kube-api-access-vk9kx") pod "7a3d6a4a-3683-4036-8171-b6488fc6b217" (UID: "7a3d6a4a-3683-4036-8171-b6488fc6b217"). InnerVolumeSpecName "kube-api-access-vk9kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.683386 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a3d6a4a-3683-4036-8171-b6488fc6b217" (UID: "7a3d6a4a-3683-4036-8171-b6488fc6b217"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.723856 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a3d6a4a-3683-4036-8171-b6488fc6b217-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.723908 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk9kx\" (UniqueName: \"kubernetes.io/projected/7a3d6a4a-3683-4036-8171-b6488fc6b217-kube-api-access-vk9kx\") on node \"crc\" DevicePath \"\"" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.828715 4767 generic.go:334] "Generic (PLEG): container finished" podID="7a3d6a4a-3683-4036-8171-b6488fc6b217" containerID="31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd" exitCode=0 Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.828798 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dsjwf" event={"ID":"7a3d6a4a-3683-4036-8171-b6488fc6b217","Type":"ContainerDied","Data":"31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd"} Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.828843 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dsjwf" event={"ID":"7a3d6a4a-3683-4036-8171-b6488fc6b217","Type":"ContainerDied","Data":"7f00cbc6d4c00b82838315af6b66de59c80fc66a75529c8493a06c592e9f9aed"} Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.828840 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dsjwf" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.828863 4767 scope.go:117] "RemoveContainer" containerID="31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.832154 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2jhp" event={"ID":"5dbf9e0d-191f-4b28-be16-19a915eb94e4","Type":"ContainerStarted","Data":"58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e"} Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.864631 4767 scope.go:117] "RemoveContainer" containerID="f92eb9a109ebf9102ec71a27648137d8e8b5fa2b290dc68109614b9dc94dfbff" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.879146 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l2jhp" podStartSLOduration=2.183999496 podStartE2EDuration="5.879110682s" podCreationTimestamp="2026-03-17 16:35:03 +0000 UTC" firstStartedPulling="2026-03-17 16:35:04.763436777 +0000 UTC m=+3496.176752824" lastFinishedPulling="2026-03-17 16:35:08.458547963 +0000 UTC m=+3499.871864010" observedRunningTime="2026-03-17 16:35:08.855500208 +0000 UTC m=+3500.268816255" watchObservedRunningTime="2026-03-17 16:35:08.879110682 +0000 UTC m=+3500.292426729" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.895134 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dsjwf"] Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.900235 4767 scope.go:117] "RemoveContainer" containerID="7ec299dce77cb11f7c42ff2808ec238639cce0d064504cca4f2b55430367a5ea" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.915423 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dsjwf"] Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.967876 4767 scope.go:117] "RemoveContainer" containerID="31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd" Mar 17 16:35:08 crc kubenswrapper[4767]: E0317 16:35:08.968544 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd\": container with ID starting with 31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd not found: ID does not exist" containerID="31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.968577 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd"} err="failed to get container status \"31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd\": rpc error: code = NotFound desc = could not find container \"31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd\": container with ID starting with 31f827f39ec6baa63d5b033bb789470c5983cebad20358b2e57a6fba4ccf52dd not found: ID does not exist" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.968601 4767 scope.go:117] "RemoveContainer" containerID="f92eb9a109ebf9102ec71a27648137d8e8b5fa2b290dc68109614b9dc94dfbff" Mar 17 16:35:08 crc kubenswrapper[4767]: E0317 16:35:08.970605 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f92eb9a109ebf9102ec71a27648137d8e8b5fa2b290dc68109614b9dc94dfbff\": container with ID starting with f92eb9a109ebf9102ec71a27648137d8e8b5fa2b290dc68109614b9dc94dfbff not found: ID does not exist" containerID="f92eb9a109ebf9102ec71a27648137d8e8b5fa2b290dc68109614b9dc94dfbff" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.970648 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92eb9a109ebf9102ec71a27648137d8e8b5fa2b290dc68109614b9dc94dfbff"} err="failed to get container status \"f92eb9a109ebf9102ec71a27648137d8e8b5fa2b290dc68109614b9dc94dfbff\": rpc error: code = NotFound desc = could not find container \"f92eb9a109ebf9102ec71a27648137d8e8b5fa2b290dc68109614b9dc94dfbff\": container with ID starting with f92eb9a109ebf9102ec71a27648137d8e8b5fa2b290dc68109614b9dc94dfbff not found: ID does not exist" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.970676 4767 scope.go:117] "RemoveContainer" containerID="7ec299dce77cb11f7c42ff2808ec238639cce0d064504cca4f2b55430367a5ea" Mar 17 16:35:08 crc kubenswrapper[4767]: E0317 16:35:08.971390 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ec299dce77cb11f7c42ff2808ec238639cce0d064504cca4f2b55430367a5ea\": container with ID starting with 7ec299dce77cb11f7c42ff2808ec238639cce0d064504cca4f2b55430367a5ea not found: ID does not exist" containerID="7ec299dce77cb11f7c42ff2808ec238639cce0d064504cca4f2b55430367a5ea" Mar 17 16:35:08 crc kubenswrapper[4767]: I0317 16:35:08.971417 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ec299dce77cb11f7c42ff2808ec238639cce0d064504cca4f2b55430367a5ea"} err="failed to get container status \"7ec299dce77cb11f7c42ff2808ec238639cce0d064504cca4f2b55430367a5ea\": rpc error: code = NotFound desc = could not find container \"7ec299dce77cb11f7c42ff2808ec238639cce0d064504cca4f2b55430367a5ea\": container with ID starting with 7ec299dce77cb11f7c42ff2808ec238639cce0d064504cca4f2b55430367a5ea not found: ID does not exist" Mar 17 16:35:09 crc kubenswrapper[4767]: I0317 16:35:09.374014 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a3d6a4a-3683-4036-8171-b6488fc6b217" path="/var/lib/kubelet/pods/7a3d6a4a-3683-4036-8171-b6488fc6b217/volumes" Mar 17 16:35:13 crc kubenswrapper[4767]: I0317 16:35:13.521524 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:13 crc kubenswrapper[4767]: I0317 16:35:13.522395 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:13 crc kubenswrapper[4767]: I0317 16:35:13.575239 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:13 crc kubenswrapper[4767]: I0317 16:35:13.947681 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:15 crc kubenswrapper[4767]: I0317 16:35:15.772582 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2jhp"] Mar 17 16:35:15 crc kubenswrapper[4767]: I0317 16:35:15.916020 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l2jhp" podUID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" containerName="registry-server" containerID="cri-o://58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e" gracePeriod=2 Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.447462 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.544990 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-catalog-content\") pod \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.545226 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk2p4\" (UniqueName: \"kubernetes.io/projected/5dbf9e0d-191f-4b28-be16-19a915eb94e4-kube-api-access-dk2p4\") pod \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.545316 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-utilities\") pod \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\" (UID: \"5dbf9e0d-191f-4b28-be16-19a915eb94e4\") " Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.547130 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-utilities" (OuterVolumeSpecName: "utilities") pod "5dbf9e0d-191f-4b28-be16-19a915eb94e4" (UID: "5dbf9e0d-191f-4b28-be16-19a915eb94e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.551522 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dbf9e0d-191f-4b28-be16-19a915eb94e4-kube-api-access-dk2p4" (OuterVolumeSpecName: "kube-api-access-dk2p4") pod "5dbf9e0d-191f-4b28-be16-19a915eb94e4" (UID: "5dbf9e0d-191f-4b28-be16-19a915eb94e4"). InnerVolumeSpecName "kube-api-access-dk2p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.574049 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5dbf9e0d-191f-4b28-be16-19a915eb94e4" (UID: "5dbf9e0d-191f-4b28-be16-19a915eb94e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.648250 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.648295 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk2p4\" (UniqueName: \"kubernetes.io/projected/5dbf9e0d-191f-4b28-be16-19a915eb94e4-kube-api-access-dk2p4\") on node \"crc\" DevicePath \"\"" Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.648308 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbf9e0d-191f-4b28-be16-19a915eb94e4-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.929206 4767 generic.go:334] "Generic (PLEG): container finished" podID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" containerID="58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e" exitCode=0 Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.929253 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2jhp" event={"ID":"5dbf9e0d-191f-4b28-be16-19a915eb94e4","Type":"ContainerDied","Data":"58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e"} Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.929287 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2jhp" event={"ID":"5dbf9e0d-191f-4b28-be16-19a915eb94e4","Type":"ContainerDied","Data":"ea9d5723cf82270bb2632b50406a82e3f21f7df76330b4f90793967691a11c6c"} Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.929295 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2jhp" Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.929310 4767 scope.go:117] "RemoveContainer" containerID="58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e" Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.964779 4767 scope.go:117] "RemoveContainer" containerID="ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081" Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.979764 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2jhp"] Mar 17 16:35:16 crc kubenswrapper[4767]: I0317 16:35:16.988634 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2jhp"] Mar 17 16:35:17 crc kubenswrapper[4767]: I0317 16:35:17.004809 4767 scope.go:117] "RemoveContainer" containerID="5b9c96e0bcc9e06dcc47ceec01e9d59647c2d284829492aebd7cbd7966234e5a" Mar 17 16:35:17 crc kubenswrapper[4767]: I0317 16:35:17.059919 4767 scope.go:117] "RemoveContainer" containerID="58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e" Mar 17 16:35:17 crc kubenswrapper[4767]: E0317 16:35:17.060448 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e\": container with ID starting with 58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e not found: ID does not exist" containerID="58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e" Mar 17 16:35:17 crc kubenswrapper[4767]: I0317 16:35:17.060495 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e"} err="failed to get container status \"58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e\": rpc error: code = NotFound desc = could not find container \"58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e\": container with ID starting with 58be2af60fd85f8cbe244d1761520dfeea67de6e03f72983cfe05552bdde128e not found: ID does not exist" Mar 17 16:35:17 crc kubenswrapper[4767]: I0317 16:35:17.060530 4767 scope.go:117] "RemoveContainer" containerID="ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081" Mar 17 16:35:17 crc kubenswrapper[4767]: E0317 16:35:17.060946 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081\": container with ID starting with ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081 not found: ID does not exist" containerID="ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081" Mar 17 16:35:17 crc kubenswrapper[4767]: I0317 16:35:17.060993 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081"} err="failed to get container status \"ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081\": rpc error: code = NotFound desc = could not find container \"ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081\": container with ID starting with ada320782bf27cc0611482d41590d6b09630937347b49591d23290c4c99be081 not found: ID does not exist" Mar 17 16:35:17 crc kubenswrapper[4767]: I0317 16:35:17.061029 4767 scope.go:117] "RemoveContainer" containerID="5b9c96e0bcc9e06dcc47ceec01e9d59647c2d284829492aebd7cbd7966234e5a" Mar 17 16:35:17 crc kubenswrapper[4767]: E0317 16:35:17.061509 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b9c96e0bcc9e06dcc47ceec01e9d59647c2d284829492aebd7cbd7966234e5a\": container with ID starting with 5b9c96e0bcc9e06dcc47ceec01e9d59647c2d284829492aebd7cbd7966234e5a not found: ID does not exist" containerID="5b9c96e0bcc9e06dcc47ceec01e9d59647c2d284829492aebd7cbd7966234e5a" Mar 17 16:35:17 crc kubenswrapper[4767]: I0317 16:35:17.061540 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b9c96e0bcc9e06dcc47ceec01e9d59647c2d284829492aebd7cbd7966234e5a"} err="failed to get container status \"5b9c96e0bcc9e06dcc47ceec01e9d59647c2d284829492aebd7cbd7966234e5a\": rpc error: code = NotFound desc = could not find container \"5b9c96e0bcc9e06dcc47ceec01e9d59647c2d284829492aebd7cbd7966234e5a\": container with ID starting with 5b9c96e0bcc9e06dcc47ceec01e9d59647c2d284829492aebd7cbd7966234e5a not found: ID does not exist" Mar 17 16:35:17 crc kubenswrapper[4767]: I0317 16:35:17.369664 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" path="/var/lib/kubelet/pods/5dbf9e0d-191f-4b28-be16-19a915eb94e4/volumes" Mar 17 16:35:34 crc kubenswrapper[4767]: I0317 16:35:34.166895 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:35:34 crc kubenswrapper[4767]: I0317 16:35:34.168483 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.155288 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562756-xqfgt"] Mar 17 16:36:00 crc kubenswrapper[4767]: E0317 16:36:00.156635 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a3d6a4a-3683-4036-8171-b6488fc6b217" containerName="extract-utilities" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.156657 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a3d6a4a-3683-4036-8171-b6488fc6b217" containerName="extract-utilities" Mar 17 16:36:00 crc kubenswrapper[4767]: E0317 16:36:00.156688 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a3d6a4a-3683-4036-8171-b6488fc6b217" containerName="extract-content" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.156696 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a3d6a4a-3683-4036-8171-b6488fc6b217" containerName="extract-content" Mar 17 16:36:00 crc kubenswrapper[4767]: E0317 16:36:00.156714 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" containerName="extract-content" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.156722 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" containerName="extract-content" Mar 17 16:36:00 crc kubenswrapper[4767]: E0317 16:36:00.156764 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a3d6a4a-3683-4036-8171-b6488fc6b217" containerName="registry-server" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.156774 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a3d6a4a-3683-4036-8171-b6488fc6b217" containerName="registry-server" Mar 17 16:36:00 crc kubenswrapper[4767]: E0317 16:36:00.156803 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" containerName="registry-server" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.156810 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" containerName="registry-server" Mar 17 16:36:00 crc kubenswrapper[4767]: E0317 16:36:00.156826 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" containerName="extract-utilities" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.156834 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" containerName="extract-utilities" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.157085 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a3d6a4a-3683-4036-8171-b6488fc6b217" containerName="registry-server" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.157102 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbf9e0d-191f-4b28-be16-19a915eb94e4" containerName="registry-server" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.158385 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562756-xqfgt" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.161761 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.161870 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.163137 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.171885 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562756-xqfgt"] Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.226041 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6hzk\" (UniqueName: \"kubernetes.io/projected/a627c42e-14c6-4cfa-9a54-e90330e5bac1-kube-api-access-c6hzk\") pod \"auto-csr-approver-29562756-xqfgt\" (UID: \"a627c42e-14c6-4cfa-9a54-e90330e5bac1\") " pod="openshift-infra/auto-csr-approver-29562756-xqfgt" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.329414 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6hzk\" (UniqueName: \"kubernetes.io/projected/a627c42e-14c6-4cfa-9a54-e90330e5bac1-kube-api-access-c6hzk\") pod \"auto-csr-approver-29562756-xqfgt\" (UID: \"a627c42e-14c6-4cfa-9a54-e90330e5bac1\") " pod="openshift-infra/auto-csr-approver-29562756-xqfgt" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.370900 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6hzk\" (UniqueName: \"kubernetes.io/projected/a627c42e-14c6-4cfa-9a54-e90330e5bac1-kube-api-access-c6hzk\") pod \"auto-csr-approver-29562756-xqfgt\" (UID: \"a627c42e-14c6-4cfa-9a54-e90330e5bac1\") " pod="openshift-infra/auto-csr-approver-29562756-xqfgt" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.481608 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562756-xqfgt" Mar 17 16:36:00 crc kubenswrapper[4767]: I0317 16:36:00.998551 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562756-xqfgt"] Mar 17 16:36:01 crc kubenswrapper[4767]: I0317 16:36:01.732790 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562756-xqfgt" event={"ID":"a627c42e-14c6-4cfa-9a54-e90330e5bac1","Type":"ContainerStarted","Data":"04508adb0cfbc2aede7bcc684dd70773f86825990efeb5c8ca286dd7bd3456ff"} Mar 17 16:36:02 crc kubenswrapper[4767]: I0317 16:36:02.757102 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562756-xqfgt" event={"ID":"a627c42e-14c6-4cfa-9a54-e90330e5bac1","Type":"ContainerStarted","Data":"1ee47c58c8afecab1a8411d37f8156529455b679c519347970ae952250374dd1"} Mar 17 16:36:02 crc kubenswrapper[4767]: I0317 16:36:02.787707 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562756-xqfgt" podStartSLOduration=1.748306165 podStartE2EDuration="2.78768s" podCreationTimestamp="2026-03-17 16:36:00 +0000 UTC" firstStartedPulling="2026-03-17 16:36:01.001207362 +0000 UTC m=+3552.414523429" lastFinishedPulling="2026-03-17 16:36:02.040581217 +0000 UTC m=+3553.453897264" observedRunningTime="2026-03-17 16:36:02.779314888 +0000 UTC m=+3554.192630925" watchObservedRunningTime="2026-03-17 16:36:02.78768 +0000 UTC m=+3554.200996057" Mar 17 16:36:03 crc kubenswrapper[4767]: I0317 16:36:03.769557 4767 generic.go:334] "Generic (PLEG): container finished" podID="a627c42e-14c6-4cfa-9a54-e90330e5bac1" containerID="1ee47c58c8afecab1a8411d37f8156529455b679c519347970ae952250374dd1" exitCode=0 Mar 17 16:36:03 crc kubenswrapper[4767]: I0317 16:36:03.769675 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562756-xqfgt" event={"ID":"a627c42e-14c6-4cfa-9a54-e90330e5bac1","Type":"ContainerDied","Data":"1ee47c58c8afecab1a8411d37f8156529455b679c519347970ae952250374dd1"} Mar 17 16:36:04 crc kubenswrapper[4767]: I0317 16:36:04.166210 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:36:04 crc kubenswrapper[4767]: I0317 16:36:04.166583 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:36:04 crc kubenswrapper[4767]: I0317 16:36:04.166646 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 16:36:04 crc kubenswrapper[4767]: I0317 16:36:04.167967 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a2d2cded09e1efc4abc6d103ad883c0499602c9a3b45371df763f0966d84f955"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 16:36:04 crc kubenswrapper[4767]: I0317 16:36:04.168050 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://a2d2cded09e1efc4abc6d103ad883c0499602c9a3b45371df763f0966d84f955" gracePeriod=600 Mar 17 16:36:04 crc kubenswrapper[4767]: I0317 16:36:04.787786 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="a2d2cded09e1efc4abc6d103ad883c0499602c9a3b45371df763f0966d84f955" exitCode=0 Mar 17 16:36:04 crc kubenswrapper[4767]: I0317 16:36:04.787848 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"a2d2cded09e1efc4abc6d103ad883c0499602c9a3b45371df763f0966d84f955"} Mar 17 16:36:04 crc kubenswrapper[4767]: I0317 16:36:04.788284 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055"} Mar 17 16:36:04 crc kubenswrapper[4767]: I0317 16:36:04.788310 4767 scope.go:117] "RemoveContainer" containerID="343630ab081828e765777f3101772f103a32ef9af618be35ed6febca6702d1ad" Mar 17 16:36:05 crc kubenswrapper[4767]: I0317 16:36:05.372794 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562756-xqfgt" Mar 17 16:36:05 crc kubenswrapper[4767]: I0317 16:36:05.488259 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6hzk\" (UniqueName: \"kubernetes.io/projected/a627c42e-14c6-4cfa-9a54-e90330e5bac1-kube-api-access-c6hzk\") pod \"a627c42e-14c6-4cfa-9a54-e90330e5bac1\" (UID: \"a627c42e-14c6-4cfa-9a54-e90330e5bac1\") " Mar 17 16:36:05 crc kubenswrapper[4767]: I0317 16:36:05.495397 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a627c42e-14c6-4cfa-9a54-e90330e5bac1-kube-api-access-c6hzk" (OuterVolumeSpecName: "kube-api-access-c6hzk") pod "a627c42e-14c6-4cfa-9a54-e90330e5bac1" (UID: "a627c42e-14c6-4cfa-9a54-e90330e5bac1"). InnerVolumeSpecName "kube-api-access-c6hzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:36:05 crc kubenswrapper[4767]: I0317 16:36:05.592741 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6hzk\" (UniqueName: \"kubernetes.io/projected/a627c42e-14c6-4cfa-9a54-e90330e5bac1-kube-api-access-c6hzk\") on node \"crc\" DevicePath \"\"" Mar 17 16:36:05 crc kubenswrapper[4767]: I0317 16:36:05.836042 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562756-xqfgt" event={"ID":"a627c42e-14c6-4cfa-9a54-e90330e5bac1","Type":"ContainerDied","Data":"04508adb0cfbc2aede7bcc684dd70773f86825990efeb5c8ca286dd7bd3456ff"} Mar 17 16:36:05 crc kubenswrapper[4767]: I0317 16:36:05.836100 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04508adb0cfbc2aede7bcc684dd70773f86825990efeb5c8ca286dd7bd3456ff" Mar 17 16:36:05 crc kubenswrapper[4767]: I0317 16:36:05.836396 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562756-xqfgt" Mar 17 16:36:05 crc kubenswrapper[4767]: I0317 16:36:05.906552 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562750-nnp6x"] Mar 17 16:36:05 crc kubenswrapper[4767]: I0317 16:36:05.937374 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562750-nnp6x"] Mar 17 16:36:07 crc kubenswrapper[4767]: I0317 16:36:07.369870 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="000288b2-7636-4225-8278-6af17a49cc09" path="/var/lib/kubelet/pods/000288b2-7636-4225-8278-6af17a49cc09/volumes" Mar 17 16:36:22 crc kubenswrapper[4767]: I0317 16:36:22.388425 4767 scope.go:117] "RemoveContainer" containerID="dd982ed329f3980d9a7abefd5211f50bf7815c10156af2e08feaa477b634d16e" Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.157743 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562758-4tl9x"] Mar 17 16:38:00 crc kubenswrapper[4767]: E0317 16:38:00.161299 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a627c42e-14c6-4cfa-9a54-e90330e5bac1" containerName="oc" Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.161349 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a627c42e-14c6-4cfa-9a54-e90330e5bac1" containerName="oc" Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.161754 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a627c42e-14c6-4cfa-9a54-e90330e5bac1" containerName="oc" Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.163244 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562758-4tl9x" Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.167633 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.167933 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.168098 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.170910 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7km5\" (UniqueName: \"kubernetes.io/projected/ebf78eae-e983-4e09-bd51-12e228f346b2-kube-api-access-h7km5\") pod \"auto-csr-approver-29562758-4tl9x\" (UID: \"ebf78eae-e983-4e09-bd51-12e228f346b2\") " pod="openshift-infra/auto-csr-approver-29562758-4tl9x" Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.172816 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562758-4tl9x"] Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.273921 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7km5\" (UniqueName: \"kubernetes.io/projected/ebf78eae-e983-4e09-bd51-12e228f346b2-kube-api-access-h7km5\") pod \"auto-csr-approver-29562758-4tl9x\" (UID: \"ebf78eae-e983-4e09-bd51-12e228f346b2\") " pod="openshift-infra/auto-csr-approver-29562758-4tl9x" Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.294452 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7km5\" (UniqueName: \"kubernetes.io/projected/ebf78eae-e983-4e09-bd51-12e228f346b2-kube-api-access-h7km5\") pod \"auto-csr-approver-29562758-4tl9x\" (UID: \"ebf78eae-e983-4e09-bd51-12e228f346b2\") " pod="openshift-infra/auto-csr-approver-29562758-4tl9x" Mar 17 16:38:00 crc kubenswrapper[4767]: I0317 16:38:00.521008 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562758-4tl9x" Mar 17 16:38:01 crc kubenswrapper[4767]: I0317 16:38:01.052270 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562758-4tl9x"] Mar 17 16:38:01 crc kubenswrapper[4767]: I0317 16:38:01.652850 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562758-4tl9x" event={"ID":"ebf78eae-e983-4e09-bd51-12e228f346b2","Type":"ContainerStarted","Data":"a0abf8cc552940e4c8c5cee02b53b1a3c1b7ec1dbfc3f81d0461e0c9ce25bf90"} Mar 17 16:38:03 crc kubenswrapper[4767]: I0317 16:38:03.676432 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562758-4tl9x" event={"ID":"ebf78eae-e983-4e09-bd51-12e228f346b2","Type":"ContainerStarted","Data":"fd32add505bf548bfc0b48e1bd3b80ce42883a993187c7c25e876d6e186534aa"} Mar 17 16:38:03 crc kubenswrapper[4767]: I0317 16:38:03.694864 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562758-4tl9x" podStartSLOduration=1.569758386 podStartE2EDuration="3.694840517s" podCreationTimestamp="2026-03-17 16:38:00 +0000 UTC" firstStartedPulling="2026-03-17 16:38:01.057045956 +0000 UTC m=+3672.470362003" lastFinishedPulling="2026-03-17 16:38:03.182128087 +0000 UTC m=+3674.595444134" observedRunningTime="2026-03-17 16:38:03.691331255 +0000 UTC m=+3675.104647312" watchObservedRunningTime="2026-03-17 16:38:03.694840517 +0000 UTC m=+3675.108156564" Mar 17 16:38:04 crc kubenswrapper[4767]: I0317 16:38:04.166973 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:38:04 crc kubenswrapper[4767]: I0317 16:38:04.167086 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:38:04 crc kubenswrapper[4767]: I0317 16:38:04.694329 4767 generic.go:334] "Generic (PLEG): container finished" podID="ebf78eae-e983-4e09-bd51-12e228f346b2" containerID="fd32add505bf548bfc0b48e1bd3b80ce42883a993187c7c25e876d6e186534aa" exitCode=0 Mar 17 16:38:04 crc kubenswrapper[4767]: I0317 16:38:04.694380 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562758-4tl9x" event={"ID":"ebf78eae-e983-4e09-bd51-12e228f346b2","Type":"ContainerDied","Data":"fd32add505bf548bfc0b48e1bd3b80ce42883a993187c7c25e876d6e186534aa"} Mar 17 16:38:06 crc kubenswrapper[4767]: I0317 16:38:06.433827 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562758-4tl9x" Mar 17 16:38:06 crc kubenswrapper[4767]: I0317 16:38:06.614323 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7km5\" (UniqueName: \"kubernetes.io/projected/ebf78eae-e983-4e09-bd51-12e228f346b2-kube-api-access-h7km5\") pod \"ebf78eae-e983-4e09-bd51-12e228f346b2\" (UID: \"ebf78eae-e983-4e09-bd51-12e228f346b2\") " Mar 17 16:38:06 crc kubenswrapper[4767]: I0317 16:38:06.622886 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebf78eae-e983-4e09-bd51-12e228f346b2-kube-api-access-h7km5" (OuterVolumeSpecName: "kube-api-access-h7km5") pod "ebf78eae-e983-4e09-bd51-12e228f346b2" (UID: "ebf78eae-e983-4e09-bd51-12e228f346b2"). InnerVolumeSpecName "kube-api-access-h7km5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:38:06 crc kubenswrapper[4767]: I0317 16:38:06.725705 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562758-4tl9x" event={"ID":"ebf78eae-e983-4e09-bd51-12e228f346b2","Type":"ContainerDied","Data":"a0abf8cc552940e4c8c5cee02b53b1a3c1b7ec1dbfc3f81d0461e0c9ce25bf90"} Mar 17 16:38:06 crc kubenswrapper[4767]: I0317 16:38:06.725761 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0abf8cc552940e4c8c5cee02b53b1a3c1b7ec1dbfc3f81d0461e0c9ce25bf90" Mar 17 16:38:06 crc kubenswrapper[4767]: I0317 16:38:06.725842 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562758-4tl9x" Mar 17 16:38:06 crc kubenswrapper[4767]: I0317 16:38:06.741220 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7km5\" (UniqueName: \"kubernetes.io/projected/ebf78eae-e983-4e09-bd51-12e228f346b2-kube-api-access-h7km5\") on node \"crc\" DevicePath \"\"" Mar 17 16:38:06 crc kubenswrapper[4767]: I0317 16:38:06.787957 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562752-98tkl"] Mar 17 16:38:06 crc kubenswrapper[4767]: I0317 16:38:06.804917 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562752-98tkl"] Mar 17 16:38:07 crc kubenswrapper[4767]: I0317 16:38:07.368794 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc79c7ba-34fb-4553-a163-3a1bfd0c8738" path="/var/lib/kubelet/pods/fc79c7ba-34fb-4553-a163-3a1bfd0c8738/volumes" Mar 17 16:38:22 crc kubenswrapper[4767]: I0317 16:38:22.532122 4767 scope.go:117] "RemoveContainer" containerID="cf5925bd611fef08bd333b4cec5746d486a74b8de1d6a9291d54393e8bc8a447" Mar 17 16:38:34 crc kubenswrapper[4767]: I0317 16:38:34.167019 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:38:34 crc kubenswrapper[4767]: I0317 16:38:34.167709 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:39:04 crc kubenswrapper[4767]: I0317 16:39:04.166813 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:39:04 crc kubenswrapper[4767]: I0317 16:39:04.167471 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:39:04 crc kubenswrapper[4767]: I0317 16:39:04.167529 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 16:39:04 crc kubenswrapper[4767]: I0317 16:39:04.168731 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 16:39:04 crc kubenswrapper[4767]: I0317 16:39:04.168795 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" gracePeriod=600 Mar 17 16:39:04 crc kubenswrapper[4767]: E0317 16:39:04.295220 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:39:04 crc kubenswrapper[4767]: I0317 16:39:04.680885 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" exitCode=0 Mar 17 16:39:04 crc kubenswrapper[4767]: I0317 16:39:04.680951 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055"} Mar 17 16:39:04 crc kubenswrapper[4767]: I0317 16:39:04.681033 4767 scope.go:117] "RemoveContainer" containerID="a2d2cded09e1efc4abc6d103ad883c0499602c9a3b45371df763f0966d84f955" Mar 17 16:39:04 crc kubenswrapper[4767]: I0317 16:39:04.682199 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:39:04 crc kubenswrapper[4767]: E0317 16:39:04.682615 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:39:19 crc kubenswrapper[4767]: I0317 16:39:19.364008 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:39:19 crc kubenswrapper[4767]: E0317 16:39:19.365051 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:39:34 crc kubenswrapper[4767]: I0317 16:39:34.354942 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:39:34 crc kubenswrapper[4767]: E0317 16:39:34.355918 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:39:45 crc kubenswrapper[4767]: I0317 16:39:45.355552 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:39:45 crc kubenswrapper[4767]: E0317 16:39:45.356455 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:39:59 crc kubenswrapper[4767]: I0317 16:39:59.365827 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:39:59 crc kubenswrapper[4767]: E0317 16:39:59.367045 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.170671 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562760-k2rwn"] Mar 17 16:40:00 crc kubenswrapper[4767]: E0317 16:40:00.171904 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf78eae-e983-4e09-bd51-12e228f346b2" containerName="oc" Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.171931 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf78eae-e983-4e09-bd51-12e228f346b2" containerName="oc" Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.172307 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf78eae-e983-4e09-bd51-12e228f346b2" containerName="oc" Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.173530 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562760-k2rwn" Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.178545 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.179086 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.179964 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.188966 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562760-k2rwn"] Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.214773 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsxtm\" (UniqueName: \"kubernetes.io/projected/4b41df7a-d8a2-4883-8018-f306322bc3da-kube-api-access-bsxtm\") pod \"auto-csr-approver-29562760-k2rwn\" (UID: \"4b41df7a-d8a2-4883-8018-f306322bc3da\") " pod="openshift-infra/auto-csr-approver-29562760-k2rwn" Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.317867 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsxtm\" (UniqueName: \"kubernetes.io/projected/4b41df7a-d8a2-4883-8018-f306322bc3da-kube-api-access-bsxtm\") pod \"auto-csr-approver-29562760-k2rwn\" (UID: \"4b41df7a-d8a2-4883-8018-f306322bc3da\") " pod="openshift-infra/auto-csr-approver-29562760-k2rwn" Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.338814 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsxtm\" (UniqueName: \"kubernetes.io/projected/4b41df7a-d8a2-4883-8018-f306322bc3da-kube-api-access-bsxtm\") pod \"auto-csr-approver-29562760-k2rwn\" (UID: \"4b41df7a-d8a2-4883-8018-f306322bc3da\") " pod="openshift-infra/auto-csr-approver-29562760-k2rwn" Mar 17 16:40:00 crc kubenswrapper[4767]: I0317 16:40:00.514340 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562760-k2rwn" Mar 17 16:40:01 crc kubenswrapper[4767]: I0317 16:40:01.069118 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562760-k2rwn"] Mar 17 16:40:01 crc kubenswrapper[4767]: I0317 16:40:01.081508 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 16:40:01 crc kubenswrapper[4767]: I0317 16:40:01.710415 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562760-k2rwn" event={"ID":"4b41df7a-d8a2-4883-8018-f306322bc3da","Type":"ContainerStarted","Data":"a757a63037f76bdfb0544a9b054e0a3e13c6dd574c2457da3cee9ce36d1406d5"} Mar 17 16:40:03 crc kubenswrapper[4767]: I0317 16:40:03.753004 4767 generic.go:334] "Generic (PLEG): container finished" podID="4b41df7a-d8a2-4883-8018-f306322bc3da" containerID="9c4929e311dc15005520721e20da67c4f563c8416bc13c29c5104776f62d5ea8" exitCode=0 Mar 17 16:40:03 crc kubenswrapper[4767]: I0317 16:40:03.753104 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562760-k2rwn" event={"ID":"4b41df7a-d8a2-4883-8018-f306322bc3da","Type":"ContainerDied","Data":"9c4929e311dc15005520721e20da67c4f563c8416bc13c29c5104776f62d5ea8"} Mar 17 16:40:05 crc kubenswrapper[4767]: I0317 16:40:05.212203 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562760-k2rwn" Mar 17 16:40:05 crc kubenswrapper[4767]: I0317 16:40:05.289270 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsxtm\" (UniqueName: \"kubernetes.io/projected/4b41df7a-d8a2-4883-8018-f306322bc3da-kube-api-access-bsxtm\") pod \"4b41df7a-d8a2-4883-8018-f306322bc3da\" (UID: \"4b41df7a-d8a2-4883-8018-f306322bc3da\") " Mar 17 16:40:05 crc kubenswrapper[4767]: I0317 16:40:05.297410 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b41df7a-d8a2-4883-8018-f306322bc3da-kube-api-access-bsxtm" (OuterVolumeSpecName: "kube-api-access-bsxtm") pod "4b41df7a-d8a2-4883-8018-f306322bc3da" (UID: "4b41df7a-d8a2-4883-8018-f306322bc3da"). InnerVolumeSpecName "kube-api-access-bsxtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:40:05 crc kubenswrapper[4767]: I0317 16:40:05.396414 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsxtm\" (UniqueName: \"kubernetes.io/projected/4b41df7a-d8a2-4883-8018-f306322bc3da-kube-api-access-bsxtm\") on node \"crc\" DevicePath \"\"" Mar 17 16:40:05 crc kubenswrapper[4767]: I0317 16:40:05.778325 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562760-k2rwn" event={"ID":"4b41df7a-d8a2-4883-8018-f306322bc3da","Type":"ContainerDied","Data":"a757a63037f76bdfb0544a9b054e0a3e13c6dd574c2457da3cee9ce36d1406d5"} Mar 17 16:40:05 crc kubenswrapper[4767]: I0317 16:40:05.778609 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a757a63037f76bdfb0544a9b054e0a3e13c6dd574c2457da3cee9ce36d1406d5" Mar 17 16:40:05 crc kubenswrapper[4767]: I0317 16:40:05.778795 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562760-k2rwn" Mar 17 16:40:06 crc kubenswrapper[4767]: I0317 16:40:06.413808 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562754-mtrx5"] Mar 17 16:40:06 crc kubenswrapper[4767]: I0317 16:40:06.429338 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562754-mtrx5"] Mar 17 16:40:07 crc kubenswrapper[4767]: I0317 16:40:07.368287 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87c15745-87fb-4aff-8065-89ad1efe0f6a" path="/var/lib/kubelet/pods/87c15745-87fb-4aff-8065-89ad1efe0f6a/volumes" Mar 17 16:40:11 crc kubenswrapper[4767]: I0317 16:40:11.355041 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:40:11 crc kubenswrapper[4767]: E0317 16:40:11.356114 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:40:22 crc kubenswrapper[4767]: I0317 16:40:22.817278 4767 scope.go:117] "RemoveContainer" containerID="cbc242ab350f19501fc6801844c592b6682389281cdc667fef45c953f31febc4" Mar 17 16:40:23 crc kubenswrapper[4767]: I0317 16:40:23.356865 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:40:23 crc kubenswrapper[4767]: E0317 16:40:23.358033 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:40:34 crc kubenswrapper[4767]: I0317 16:40:34.354758 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:40:34 crc kubenswrapper[4767]: E0317 16:40:34.355891 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:40:50 crc kubenswrapper[4767]: I0317 16:40:50.366946 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:40:50 crc kubenswrapper[4767]: E0317 16:40:50.368292 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:41:01 crc kubenswrapper[4767]: I0317 16:41:01.355145 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:41:01 crc kubenswrapper[4767]: E0317 16:41:01.356148 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:41:15 crc kubenswrapper[4767]: I0317 16:41:15.355514 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:41:15 crc kubenswrapper[4767]: E0317 16:41:15.356452 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.355263 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:41:30 crc kubenswrapper[4767]: E0317 16:41:30.356236 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.747217 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pcjps"] Mar 17 16:41:30 crc kubenswrapper[4767]: E0317 16:41:30.748544 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b41df7a-d8a2-4883-8018-f306322bc3da" containerName="oc" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.748583 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b41df7a-d8a2-4883-8018-f306322bc3da" containerName="oc" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.748911 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b41df7a-d8a2-4883-8018-f306322bc3da" containerName="oc" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.751003 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.766007 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pcjps"] Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.771346 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhpjw\" (UniqueName: \"kubernetes.io/projected/6be2f750-85cd-4349-b256-4df96dc7bd42-kube-api-access-nhpjw\") pod \"redhat-operators-pcjps\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.771487 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-utilities\") pod \"redhat-operators-pcjps\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.771521 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-catalog-content\") pod \"redhat-operators-pcjps\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.873802 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhpjw\" (UniqueName: \"kubernetes.io/projected/6be2f750-85cd-4349-b256-4df96dc7bd42-kube-api-access-nhpjw\") pod \"redhat-operators-pcjps\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.874034 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-utilities\") pod \"redhat-operators-pcjps\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.874083 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-catalog-content\") pod \"redhat-operators-pcjps\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.874856 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-catalog-content\") pod \"redhat-operators-pcjps\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.874873 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-utilities\") pod \"redhat-operators-pcjps\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:30 crc kubenswrapper[4767]: I0317 16:41:30.908159 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhpjw\" (UniqueName: \"kubernetes.io/projected/6be2f750-85cd-4349-b256-4df96dc7bd42-kube-api-access-nhpjw\") pod \"redhat-operators-pcjps\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:31 crc kubenswrapper[4767]: I0317 16:41:31.081413 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:31 crc kubenswrapper[4767]: I0317 16:41:31.807773 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pcjps"] Mar 17 16:41:32 crc kubenswrapper[4767]: I0317 16:41:32.156847 4767 generic.go:334] "Generic (PLEG): container finished" podID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerID="f8871575b4a884e2e52c769808e599f02e57a3447c61bd9e4746f103b3165c0d" exitCode=0 Mar 17 16:41:32 crc kubenswrapper[4767]: I0317 16:41:32.156953 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcjps" event={"ID":"6be2f750-85cd-4349-b256-4df96dc7bd42","Type":"ContainerDied","Data":"f8871575b4a884e2e52c769808e599f02e57a3447c61bd9e4746f103b3165c0d"} Mar 17 16:41:32 crc kubenswrapper[4767]: I0317 16:41:32.157432 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcjps" event={"ID":"6be2f750-85cd-4349-b256-4df96dc7bd42","Type":"ContainerStarted","Data":"20d4d759189c47ae197a2993b8d2baf672f4fcffb1b97a4b6eec29565a0c89ff"} Mar 17 16:41:33 crc kubenswrapper[4767]: I0317 16:41:33.171018 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcjps" event={"ID":"6be2f750-85cd-4349-b256-4df96dc7bd42","Type":"ContainerStarted","Data":"5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58"} Mar 17 16:41:38 crc kubenswrapper[4767]: I0317 16:41:38.942996 4767 generic.go:334] "Generic (PLEG): container finished" podID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerID="5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58" exitCode=0 Mar 17 16:41:38 crc kubenswrapper[4767]: I0317 16:41:38.943243 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcjps" event={"ID":"6be2f750-85cd-4349-b256-4df96dc7bd42","Type":"ContainerDied","Data":"5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58"} Mar 17 16:41:39 crc kubenswrapper[4767]: I0317 16:41:39.979998 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcjps" event={"ID":"6be2f750-85cd-4349-b256-4df96dc7bd42","Type":"ContainerStarted","Data":"8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96"} Mar 17 16:41:40 crc kubenswrapper[4767]: I0317 16:41:40.046630 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pcjps" podStartSLOduration=2.771262716 podStartE2EDuration="10.046578883s" podCreationTimestamp="2026-03-17 16:41:30 +0000 UTC" firstStartedPulling="2026-03-17 16:41:32.159565168 +0000 UTC m=+3883.572881215" lastFinishedPulling="2026-03-17 16:41:39.434881335 +0000 UTC m=+3890.848197382" observedRunningTime="2026-03-17 16:41:40.020585755 +0000 UTC m=+3891.433901812" watchObservedRunningTime="2026-03-17 16:41:40.046578883 +0000 UTC m=+3891.459894930" Mar 17 16:41:41 crc kubenswrapper[4767]: I0317 16:41:41.081907 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:41 crc kubenswrapper[4767]: I0317 16:41:41.082334 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:41:42 crc kubenswrapper[4767]: I0317 16:41:42.137932 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pcjps" podUID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerName="registry-server" probeResult="failure" output=< Mar 17 16:41:42 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:41:42 crc kubenswrapper[4767]: > Mar 17 16:41:43 crc kubenswrapper[4767]: I0317 16:41:43.355462 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:41:43 crc kubenswrapper[4767]: E0317 16:41:43.355926 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:41:52 crc kubenswrapper[4767]: I0317 16:41:52.133375 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pcjps" podUID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerName="registry-server" probeResult="failure" output=< Mar 17 16:41:52 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:41:52 crc kubenswrapper[4767]: > Mar 17 16:41:55 crc kubenswrapper[4767]: I0317 16:41:55.354617 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:41:55 crc kubenswrapper[4767]: E0317 16:41:55.355503 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:42:00 crc kubenswrapper[4767]: I0317 16:42:00.168629 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562762-qmqmf"] Mar 17 16:42:00 crc kubenswrapper[4767]: I0317 16:42:00.180713 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562762-qmqmf" Mar 17 16:42:00 crc kubenswrapper[4767]: I0317 16:42:00.183659 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:42:00 crc kubenswrapper[4767]: I0317 16:42:00.183962 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:42:00 crc kubenswrapper[4767]: I0317 16:42:00.185084 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:42:00 crc kubenswrapper[4767]: I0317 16:42:00.196319 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562762-qmqmf"] Mar 17 16:42:00 crc kubenswrapper[4767]: I0317 16:42:00.314217 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9jgz\" (UniqueName: \"kubernetes.io/projected/d767c586-6134-4c00-ab0e-dd16cf2934fa-kube-api-access-d9jgz\") pod \"auto-csr-approver-29562762-qmqmf\" (UID: \"d767c586-6134-4c00-ab0e-dd16cf2934fa\") " pod="openshift-infra/auto-csr-approver-29562762-qmqmf" Mar 17 16:42:00 crc kubenswrapper[4767]: I0317 16:42:00.417825 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9jgz\" (UniqueName: \"kubernetes.io/projected/d767c586-6134-4c00-ab0e-dd16cf2934fa-kube-api-access-d9jgz\") pod \"auto-csr-approver-29562762-qmqmf\" (UID: \"d767c586-6134-4c00-ab0e-dd16cf2934fa\") " pod="openshift-infra/auto-csr-approver-29562762-qmqmf" Mar 17 16:42:00 crc kubenswrapper[4767]: I0317 16:42:00.441814 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9jgz\" (UniqueName: \"kubernetes.io/projected/d767c586-6134-4c00-ab0e-dd16cf2934fa-kube-api-access-d9jgz\") pod \"auto-csr-approver-29562762-qmqmf\" (UID: \"d767c586-6134-4c00-ab0e-dd16cf2934fa\") " pod="openshift-infra/auto-csr-approver-29562762-qmqmf" Mar 17 16:42:00 crc kubenswrapper[4767]: I0317 16:42:00.511780 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562762-qmqmf" Mar 17 16:42:01 crc kubenswrapper[4767]: I0317 16:42:01.076985 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562762-qmqmf"] Mar 17 16:42:01 crc kubenswrapper[4767]: I0317 16:42:01.145519 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:42:01 crc kubenswrapper[4767]: I0317 16:42:01.238412 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:42:01 crc kubenswrapper[4767]: I0317 16:42:01.271225 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562762-qmqmf" event={"ID":"d767c586-6134-4c00-ab0e-dd16cf2934fa","Type":"ContainerStarted","Data":"759c34364032552281a7b7f926d5b57561c8e2358e91047072531e5416e6797c"} Mar 17 16:42:02 crc kubenswrapper[4767]: I0317 16:42:02.079963 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pcjps"] Mar 17 16:42:02 crc kubenswrapper[4767]: I0317 16:42:02.286624 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pcjps" podUID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerName="registry-server" containerID="cri-o://8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96" gracePeriod=2 Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.126695 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.264668 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhpjw\" (UniqueName: \"kubernetes.io/projected/6be2f750-85cd-4349-b256-4df96dc7bd42-kube-api-access-nhpjw\") pod \"6be2f750-85cd-4349-b256-4df96dc7bd42\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.264790 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-catalog-content\") pod \"6be2f750-85cd-4349-b256-4df96dc7bd42\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.264915 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-utilities\") pod \"6be2f750-85cd-4349-b256-4df96dc7bd42\" (UID: \"6be2f750-85cd-4349-b256-4df96dc7bd42\") " Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.267817 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-utilities" (OuterVolumeSpecName: "utilities") pod "6be2f750-85cd-4349-b256-4df96dc7bd42" (UID: "6be2f750-85cd-4349-b256-4df96dc7bd42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.296211 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6be2f750-85cd-4349-b256-4df96dc7bd42-kube-api-access-nhpjw" (OuterVolumeSpecName: "kube-api-access-nhpjw") pod "6be2f750-85cd-4349-b256-4df96dc7bd42" (UID: "6be2f750-85cd-4349-b256-4df96dc7bd42"). InnerVolumeSpecName "kube-api-access-nhpjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.332152 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcjps" event={"ID":"6be2f750-85cd-4349-b256-4df96dc7bd42","Type":"ContainerDied","Data":"8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96"} Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.332270 4767 scope.go:117] "RemoveContainer" containerID="8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.332112 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcjps" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.331963 4767 generic.go:334] "Generic (PLEG): container finished" podID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerID="8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96" exitCode=0 Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.334235 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcjps" event={"ID":"6be2f750-85cd-4349-b256-4df96dc7bd42","Type":"ContainerDied","Data":"20d4d759189c47ae197a2993b8d2baf672f4fcffb1b97a4b6eec29565a0c89ff"} Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.351029 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562762-qmqmf" event={"ID":"d767c586-6134-4c00-ab0e-dd16cf2934fa","Type":"ContainerStarted","Data":"e7ff1ec459e8edbc1902c2fb66b89ddeb8e022e23ed5e41840950a15c5405291"} Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.386232 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhpjw\" (UniqueName: \"kubernetes.io/projected/6be2f750-85cd-4349-b256-4df96dc7bd42-kube-api-access-nhpjw\") on node \"crc\" DevicePath \"\"" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.386365 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.401973 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562762-qmqmf" podStartSLOduration=2.2651374459999998 podStartE2EDuration="3.401941112s" podCreationTimestamp="2026-03-17 16:42:00 +0000 UTC" firstStartedPulling="2026-03-17 16:42:01.087727841 +0000 UTC m=+3912.501043888" lastFinishedPulling="2026-03-17 16:42:02.224531507 +0000 UTC m=+3913.637847554" observedRunningTime="2026-03-17 16:42:03.39839486 +0000 UTC m=+3914.811710917" watchObservedRunningTime="2026-03-17 16:42:03.401941112 +0000 UTC m=+3914.815257169" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.443687 4767 scope.go:117] "RemoveContainer" containerID="5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.482579 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6be2f750-85cd-4349-b256-4df96dc7bd42" (UID: "6be2f750-85cd-4349-b256-4df96dc7bd42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.485012 4767 scope.go:117] "RemoveContainer" containerID="f8871575b4a884e2e52c769808e599f02e57a3447c61bd9e4746f103b3165c0d" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.488626 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6be2f750-85cd-4349-b256-4df96dc7bd42-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.541424 4767 scope.go:117] "RemoveContainer" containerID="8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96" Mar 17 16:42:03 crc kubenswrapper[4767]: E0317 16:42:03.549606 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96\": container with ID starting with 8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96 not found: ID does not exist" containerID="8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.549685 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96"} err="failed to get container status \"8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96\": rpc error: code = NotFound desc = could not find container \"8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96\": container with ID starting with 8eb52994b9fad5075f55735cb439c6f4e6cf49b8dcc777400feee1eff821db96 not found: ID does not exist" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.549717 4767 scope.go:117] "RemoveContainer" containerID="5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58" Mar 17 16:42:03 crc kubenswrapper[4767]: E0317 16:42:03.553322 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58\": container with ID starting with 5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58 not found: ID does not exist" containerID="5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.553379 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58"} err="failed to get container status \"5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58\": rpc error: code = NotFound desc = could not find container \"5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58\": container with ID starting with 5b09ea251af31bc1c8b652640465bb7b1809de01b6052fafc064235b1aabde58 not found: ID does not exist" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.553416 4767 scope.go:117] "RemoveContainer" containerID="f8871575b4a884e2e52c769808e599f02e57a3447c61bd9e4746f103b3165c0d" Mar 17 16:42:03 crc kubenswrapper[4767]: E0317 16:42:03.554652 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8871575b4a884e2e52c769808e599f02e57a3447c61bd9e4746f103b3165c0d\": container with ID starting with f8871575b4a884e2e52c769808e599f02e57a3447c61bd9e4746f103b3165c0d not found: ID does not exist" containerID="f8871575b4a884e2e52c769808e599f02e57a3447c61bd9e4746f103b3165c0d" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.554719 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8871575b4a884e2e52c769808e599f02e57a3447c61bd9e4746f103b3165c0d"} err="failed to get container status \"f8871575b4a884e2e52c769808e599f02e57a3447c61bd9e4746f103b3165c0d\": rpc error: code = NotFound desc = could not find container \"f8871575b4a884e2e52c769808e599f02e57a3447c61bd9e4746f103b3165c0d\": container with ID starting with f8871575b4a884e2e52c769808e599f02e57a3447c61bd9e4746f103b3165c0d not found: ID does not exist" Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.699098 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pcjps"] Mar 17 16:42:03 crc kubenswrapper[4767]: I0317 16:42:03.718150 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pcjps"] Mar 17 16:42:04 crc kubenswrapper[4767]: I0317 16:42:04.376190 4767 generic.go:334] "Generic (PLEG): container finished" podID="d767c586-6134-4c00-ab0e-dd16cf2934fa" containerID="e7ff1ec459e8edbc1902c2fb66b89ddeb8e022e23ed5e41840950a15c5405291" exitCode=0 Mar 17 16:42:04 crc kubenswrapper[4767]: I0317 16:42:04.376293 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562762-qmqmf" event={"ID":"d767c586-6134-4c00-ab0e-dd16cf2934fa","Type":"ContainerDied","Data":"e7ff1ec459e8edbc1902c2fb66b89ddeb8e022e23ed5e41840950a15c5405291"} Mar 17 16:42:05 crc kubenswrapper[4767]: I0317 16:42:05.371305 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6be2f750-85cd-4349-b256-4df96dc7bd42" path="/var/lib/kubelet/pods/6be2f750-85cd-4349-b256-4df96dc7bd42/volumes" Mar 17 16:42:05 crc kubenswrapper[4767]: I0317 16:42:05.866024 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562762-qmqmf" Mar 17 16:42:05 crc kubenswrapper[4767]: I0317 16:42:05.958277 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9jgz\" (UniqueName: \"kubernetes.io/projected/d767c586-6134-4c00-ab0e-dd16cf2934fa-kube-api-access-d9jgz\") pod \"d767c586-6134-4c00-ab0e-dd16cf2934fa\" (UID: \"d767c586-6134-4c00-ab0e-dd16cf2934fa\") " Mar 17 16:42:05 crc kubenswrapper[4767]: I0317 16:42:05.971128 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d767c586-6134-4c00-ab0e-dd16cf2934fa-kube-api-access-d9jgz" (OuterVolumeSpecName: "kube-api-access-d9jgz") pod "d767c586-6134-4c00-ab0e-dd16cf2934fa" (UID: "d767c586-6134-4c00-ab0e-dd16cf2934fa"). InnerVolumeSpecName "kube-api-access-d9jgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:42:06 crc kubenswrapper[4767]: I0317 16:42:06.062959 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9jgz\" (UniqueName: \"kubernetes.io/projected/d767c586-6134-4c00-ab0e-dd16cf2934fa-kube-api-access-d9jgz\") on node \"crc\" DevicePath \"\"" Mar 17 16:42:06 crc kubenswrapper[4767]: I0317 16:42:06.462941 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:42:06 crc kubenswrapper[4767]: E0317 16:42:06.464034 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:42:06 crc kubenswrapper[4767]: I0317 16:42:06.501847 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562762-qmqmf" event={"ID":"d767c586-6134-4c00-ab0e-dd16cf2934fa","Type":"ContainerDied","Data":"759c34364032552281a7b7f926d5b57561c8e2358e91047072531e5416e6797c"} Mar 17 16:42:06 crc kubenswrapper[4767]: I0317 16:42:06.501893 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="759c34364032552281a7b7f926d5b57561c8e2358e91047072531e5416e6797c" Mar 17 16:42:06 crc kubenswrapper[4767]: I0317 16:42:06.501967 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562762-qmqmf" Mar 17 16:42:06 crc kubenswrapper[4767]: I0317 16:42:06.562324 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562756-xqfgt"] Mar 17 16:42:06 crc kubenswrapper[4767]: I0317 16:42:06.580158 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562756-xqfgt"] Mar 17 16:42:07 crc kubenswrapper[4767]: I0317 16:42:07.370883 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a627c42e-14c6-4cfa-9a54-e90330e5bac1" path="/var/lib/kubelet/pods/a627c42e-14c6-4cfa-9a54-e90330e5bac1/volumes" Mar 17 16:42:17 crc kubenswrapper[4767]: I0317 16:42:17.355062 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:42:17 crc kubenswrapper[4767]: E0317 16:42:17.356316 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:42:22 crc kubenswrapper[4767]: I0317 16:42:22.950882 4767 scope.go:117] "RemoveContainer" containerID="1ee47c58c8afecab1a8411d37f8156529455b679c519347970ae952250374dd1" Mar 17 16:42:30 crc kubenswrapper[4767]: I0317 16:42:30.355199 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:42:30 crc kubenswrapper[4767]: E0317 16:42:30.356082 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:42:41 crc kubenswrapper[4767]: I0317 16:42:41.355588 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:42:41 crc kubenswrapper[4767]: E0317 16:42:41.356567 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:42:54 crc kubenswrapper[4767]: I0317 16:42:54.355577 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:42:54 crc kubenswrapper[4767]: E0317 16:42:54.356599 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:43:06 crc kubenswrapper[4767]: I0317 16:43:06.354728 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:43:06 crc kubenswrapper[4767]: E0317 16:43:06.355771 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:43:21 crc kubenswrapper[4767]: I0317 16:43:21.364012 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:43:21 crc kubenswrapper[4767]: E0317 16:43:21.365029 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:43:36 crc kubenswrapper[4767]: I0317 16:43:36.355349 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:43:36 crc kubenswrapper[4767]: E0317 16:43:36.356263 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:43:50 crc kubenswrapper[4767]: I0317 16:43:50.355335 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:43:50 crc kubenswrapper[4767]: E0317 16:43:50.356407 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.163887 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562764-zsncx"] Mar 17 16:44:00 crc kubenswrapper[4767]: E0317 16:44:00.165137 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerName="extract-utilities" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.165154 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerName="extract-utilities" Mar 17 16:44:00 crc kubenswrapper[4767]: E0317 16:44:00.165188 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerName="registry-server" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.165197 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerName="registry-server" Mar 17 16:44:00 crc kubenswrapper[4767]: E0317 16:44:00.165232 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerName="extract-content" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.165238 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerName="extract-content" Mar 17 16:44:00 crc kubenswrapper[4767]: E0317 16:44:00.165260 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d767c586-6134-4c00-ab0e-dd16cf2934fa" containerName="oc" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.165267 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d767c586-6134-4c00-ab0e-dd16cf2934fa" containerName="oc" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.165513 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d767c586-6134-4c00-ab0e-dd16cf2934fa" containerName="oc" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.165551 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="6be2f750-85cd-4349-b256-4df96dc7bd42" containerName="registry-server" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.166492 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562764-zsncx" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.168551 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.168882 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.169891 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.187457 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562764-zsncx"] Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.244447 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l74h2\" (UniqueName: \"kubernetes.io/projected/9795ce09-d4bf-46de-b861-41289e385f36-kube-api-access-l74h2\") pod \"auto-csr-approver-29562764-zsncx\" (UID: \"9795ce09-d4bf-46de-b861-41289e385f36\") " pod="openshift-infra/auto-csr-approver-29562764-zsncx" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.346318 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l74h2\" (UniqueName: \"kubernetes.io/projected/9795ce09-d4bf-46de-b861-41289e385f36-kube-api-access-l74h2\") pod \"auto-csr-approver-29562764-zsncx\" (UID: \"9795ce09-d4bf-46de-b861-41289e385f36\") " pod="openshift-infra/auto-csr-approver-29562764-zsncx" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.367258 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l74h2\" (UniqueName: \"kubernetes.io/projected/9795ce09-d4bf-46de-b861-41289e385f36-kube-api-access-l74h2\") pod \"auto-csr-approver-29562764-zsncx\" (UID: \"9795ce09-d4bf-46de-b861-41289e385f36\") " pod="openshift-infra/auto-csr-approver-29562764-zsncx" Mar 17 16:44:00 crc kubenswrapper[4767]: I0317 16:44:00.496929 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562764-zsncx" Mar 17 16:44:01 crc kubenswrapper[4767]: I0317 16:44:01.029030 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562764-zsncx"] Mar 17 16:44:01 crc kubenswrapper[4767]: I0317 16:44:01.352330 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562764-zsncx" event={"ID":"9795ce09-d4bf-46de-b861-41289e385f36","Type":"ContainerStarted","Data":"2504a6b507c9f1a5577510ee2a5485ca520e0b262e0c8956d5b1dea9473e6802"} Mar 17 16:44:02 crc kubenswrapper[4767]: I0317 16:44:02.355036 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:44:02 crc kubenswrapper[4767]: E0317 16:44:02.356210 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:44:03 crc kubenswrapper[4767]: I0317 16:44:03.386294 4767 generic.go:334] "Generic (PLEG): container finished" podID="9795ce09-d4bf-46de-b861-41289e385f36" containerID="b4e15f78acffc90e70fef8336caa63f6518d587e5458a6f42b96f535c9dbf612" exitCode=0 Mar 17 16:44:03 crc kubenswrapper[4767]: I0317 16:44:03.386586 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562764-zsncx" event={"ID":"9795ce09-d4bf-46de-b861-41289e385f36","Type":"ContainerDied","Data":"b4e15f78acffc90e70fef8336caa63f6518d587e5458a6f42b96f535c9dbf612"} Mar 17 16:44:04 crc kubenswrapper[4767]: I0317 16:44:04.865992 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562764-zsncx" Mar 17 16:44:04 crc kubenswrapper[4767]: I0317 16:44:04.929401 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l74h2\" (UniqueName: \"kubernetes.io/projected/9795ce09-d4bf-46de-b861-41289e385f36-kube-api-access-l74h2\") pod \"9795ce09-d4bf-46de-b861-41289e385f36\" (UID: \"9795ce09-d4bf-46de-b861-41289e385f36\") " Mar 17 16:44:04 crc kubenswrapper[4767]: I0317 16:44:04.944595 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9795ce09-d4bf-46de-b861-41289e385f36-kube-api-access-l74h2" (OuterVolumeSpecName: "kube-api-access-l74h2") pod "9795ce09-d4bf-46de-b861-41289e385f36" (UID: "9795ce09-d4bf-46de-b861-41289e385f36"). InnerVolumeSpecName "kube-api-access-l74h2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:44:05 crc kubenswrapper[4767]: I0317 16:44:05.036340 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l74h2\" (UniqueName: \"kubernetes.io/projected/9795ce09-d4bf-46de-b861-41289e385f36-kube-api-access-l74h2\") on node \"crc\" DevicePath \"\"" Mar 17 16:44:05 crc kubenswrapper[4767]: I0317 16:44:05.424993 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562764-zsncx" event={"ID":"9795ce09-d4bf-46de-b861-41289e385f36","Type":"ContainerDied","Data":"2504a6b507c9f1a5577510ee2a5485ca520e0b262e0c8956d5b1dea9473e6802"} Mar 17 16:44:05 crc kubenswrapper[4767]: I0317 16:44:05.425051 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2504a6b507c9f1a5577510ee2a5485ca520e0b262e0c8956d5b1dea9473e6802" Mar 17 16:44:05 crc kubenswrapper[4767]: I0317 16:44:05.425128 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562764-zsncx" Mar 17 16:44:05 crc kubenswrapper[4767]: I0317 16:44:05.952364 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562758-4tl9x"] Mar 17 16:44:05 crc kubenswrapper[4767]: I0317 16:44:05.971158 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562758-4tl9x"] Mar 17 16:44:07 crc kubenswrapper[4767]: I0317 16:44:07.371585 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebf78eae-e983-4e09-bd51-12e228f346b2" path="/var/lib/kubelet/pods/ebf78eae-e983-4e09-bd51-12e228f346b2/volumes" Mar 17 16:44:17 crc kubenswrapper[4767]: I0317 16:44:17.356454 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:44:17 crc kubenswrapper[4767]: I0317 16:44:17.919935 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"a329b73e6c74a32ee0eed732b0cfe2e5edcb9ecafe799acd643961d80496c7af"} Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.003664 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4v85c"] Mar 17 16:44:19 crc kubenswrapper[4767]: E0317 16:44:19.004832 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9795ce09-d4bf-46de-b861-41289e385f36" containerName="oc" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.004847 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="9795ce09-d4bf-46de-b861-41289e385f36" containerName="oc" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.005139 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="9795ce09-d4bf-46de-b861-41289e385f36" containerName="oc" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.007777 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.035395 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4v85c"] Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.107983 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-utilities\") pod \"certified-operators-4v85c\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.108226 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-catalog-content\") pod \"certified-operators-4v85c\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.108336 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grklw\" (UniqueName: \"kubernetes.io/projected/c3900db7-d204-4f6b-a0f3-79d172d16155-kube-api-access-grklw\") pod \"certified-operators-4v85c\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.210737 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-utilities\") pod \"certified-operators-4v85c\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.211296 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-utilities\") pod \"certified-operators-4v85c\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.211450 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-catalog-content\") pod \"certified-operators-4v85c\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.211536 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grklw\" (UniqueName: \"kubernetes.io/projected/c3900db7-d204-4f6b-a0f3-79d172d16155-kube-api-access-grklw\") pod \"certified-operators-4v85c\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.211786 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-catalog-content\") pod \"certified-operators-4v85c\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.244670 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grklw\" (UniqueName: \"kubernetes.io/projected/c3900db7-d204-4f6b-a0f3-79d172d16155-kube-api-access-grklw\") pod \"certified-operators-4v85c\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:19 crc kubenswrapper[4767]: I0317 16:44:19.360304 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:20 crc kubenswrapper[4767]: I0317 16:44:20.843242 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4v85c"] Mar 17 16:44:20 crc kubenswrapper[4767]: I0317 16:44:20.982474 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v85c" event={"ID":"c3900db7-d204-4f6b-a0f3-79d172d16155","Type":"ContainerStarted","Data":"9d4520f7afaae65f86b2e3b4b997e22a08ef84c3e5a96312e0534a3c92ef53a9"} Mar 17 16:44:21 crc kubenswrapper[4767]: I0317 16:44:21.996979 4767 generic.go:334] "Generic (PLEG): container finished" podID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerID="b23712161d77e92842a8f593accfd2933e8c360d003454d4113d7d79c7d31bbd" exitCode=0 Mar 17 16:44:21 crc kubenswrapper[4767]: I0317 16:44:21.997077 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v85c" event={"ID":"c3900db7-d204-4f6b-a0f3-79d172d16155","Type":"ContainerDied","Data":"b23712161d77e92842a8f593accfd2933e8c360d003454d4113d7d79c7d31bbd"} Mar 17 16:44:23 crc kubenswrapper[4767]: I0317 16:44:23.129244 4767 scope.go:117] "RemoveContainer" containerID="fd32add505bf548bfc0b48e1bd3b80ce42883a993187c7c25e876d6e186534aa" Mar 17 16:44:24 crc kubenswrapper[4767]: I0317 16:44:24.023365 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v85c" event={"ID":"c3900db7-d204-4f6b-a0f3-79d172d16155","Type":"ContainerStarted","Data":"5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a"} Mar 17 16:44:25 crc kubenswrapper[4767]: I0317 16:44:25.037739 4767 generic.go:334] "Generic (PLEG): container finished" podID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerID="5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a" exitCode=0 Mar 17 16:44:25 crc kubenswrapper[4767]: I0317 16:44:25.037855 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v85c" event={"ID":"c3900db7-d204-4f6b-a0f3-79d172d16155","Type":"ContainerDied","Data":"5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a"} Mar 17 16:44:26 crc kubenswrapper[4767]: I0317 16:44:26.051946 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v85c" event={"ID":"c3900db7-d204-4f6b-a0f3-79d172d16155","Type":"ContainerStarted","Data":"852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc"} Mar 17 16:44:26 crc kubenswrapper[4767]: I0317 16:44:26.075552 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4v85c" podStartSLOduration=4.5726882060000005 podStartE2EDuration="8.075526075s" podCreationTimestamp="2026-03-17 16:44:18 +0000 UTC" firstStartedPulling="2026-03-17 16:44:21.999968418 +0000 UTC m=+4053.413284465" lastFinishedPulling="2026-03-17 16:44:25.502806297 +0000 UTC m=+4056.916122334" observedRunningTime="2026-03-17 16:44:26.074779563 +0000 UTC m=+4057.488095630" watchObservedRunningTime="2026-03-17 16:44:26.075526075 +0000 UTC m=+4057.488842122" Mar 17 16:44:29 crc kubenswrapper[4767]: I0317 16:44:29.372723 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:29 crc kubenswrapper[4767]: I0317 16:44:29.373332 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:30 crc kubenswrapper[4767]: I0317 16:44:30.422701 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-4v85c" podUID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerName="registry-server" probeResult="failure" output=< Mar 17 16:44:30 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:44:30 crc kubenswrapper[4767]: > Mar 17 16:44:39 crc kubenswrapper[4767]: I0317 16:44:39.427091 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:39 crc kubenswrapper[4767]: I0317 16:44:39.487271 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:42 crc kubenswrapper[4767]: I0317 16:44:42.361016 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4v85c"] Mar 17 16:44:42 crc kubenswrapper[4767]: I0317 16:44:42.361997 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4v85c" podUID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerName="registry-server" containerID="cri-o://852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc" gracePeriod=2 Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.063716 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.270762 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-catalog-content\") pod \"c3900db7-d204-4f6b-a0f3-79d172d16155\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.270873 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grklw\" (UniqueName: \"kubernetes.io/projected/c3900db7-d204-4f6b-a0f3-79d172d16155-kube-api-access-grklw\") pod \"c3900db7-d204-4f6b-a0f3-79d172d16155\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.271316 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-utilities\") pod \"c3900db7-d204-4f6b-a0f3-79d172d16155\" (UID: \"c3900db7-d204-4f6b-a0f3-79d172d16155\") " Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.272004 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-utilities" (OuterVolumeSpecName: "utilities") pod "c3900db7-d204-4f6b-a0f3-79d172d16155" (UID: "c3900db7-d204-4f6b-a0f3-79d172d16155"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.272592 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.282905 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3900db7-d204-4f6b-a0f3-79d172d16155-kube-api-access-grklw" (OuterVolumeSpecName: "kube-api-access-grklw") pod "c3900db7-d204-4f6b-a0f3-79d172d16155" (UID: "c3900db7-d204-4f6b-a0f3-79d172d16155"). InnerVolumeSpecName "kube-api-access-grklw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.389003 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grklw\" (UniqueName: \"kubernetes.io/projected/c3900db7-d204-4f6b-a0f3-79d172d16155-kube-api-access-grklw\") on node \"crc\" DevicePath \"\"" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.402137 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3900db7-d204-4f6b-a0f3-79d172d16155" (UID: "c3900db7-d204-4f6b-a0f3-79d172d16155"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.471659 4767 generic.go:334] "Generic (PLEG): container finished" podID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerID="852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc" exitCode=0 Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.471727 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v85c" event={"ID":"c3900db7-d204-4f6b-a0f3-79d172d16155","Type":"ContainerDied","Data":"852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc"} Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.471773 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v85c" event={"ID":"c3900db7-d204-4f6b-a0f3-79d172d16155","Type":"ContainerDied","Data":"9d4520f7afaae65f86b2e3b4b997e22a08ef84c3e5a96312e0534a3c92ef53a9"} Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.471799 4767 scope.go:117] "RemoveContainer" containerID="852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.472045 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4v85c" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.498708 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3900db7-d204-4f6b-a0f3-79d172d16155-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.515198 4767 scope.go:117] "RemoveContainer" containerID="5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.538008 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4v85c"] Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.546883 4767 scope.go:117] "RemoveContainer" containerID="b23712161d77e92842a8f593accfd2933e8c360d003454d4113d7d79c7d31bbd" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.550773 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4v85c"] Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.609013 4767 scope.go:117] "RemoveContainer" containerID="852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc" Mar 17 16:44:43 crc kubenswrapper[4767]: E0317 16:44:43.609788 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc\": container with ID starting with 852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc not found: ID does not exist" containerID="852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.609827 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc"} err="failed to get container status \"852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc\": rpc error: code = NotFound desc = could not find container \"852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc\": container with ID starting with 852685873d3771544eb04ef32a578fbcc5c0ffca409af6498b2255a13ed35cfc not found: ID does not exist" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.609854 4767 scope.go:117] "RemoveContainer" containerID="5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a" Mar 17 16:44:43 crc kubenswrapper[4767]: E0317 16:44:43.610723 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a\": container with ID starting with 5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a not found: ID does not exist" containerID="5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.610840 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a"} err="failed to get container status \"5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a\": rpc error: code = NotFound desc = could not find container \"5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a\": container with ID starting with 5418a7f46cdb482814f35dc387256ee2cac7898356ce275dbc7605f7a87eb39a not found: ID does not exist" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.610935 4767 scope.go:117] "RemoveContainer" containerID="b23712161d77e92842a8f593accfd2933e8c360d003454d4113d7d79c7d31bbd" Mar 17 16:44:43 crc kubenswrapper[4767]: E0317 16:44:43.611312 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b23712161d77e92842a8f593accfd2933e8c360d003454d4113d7d79c7d31bbd\": container with ID starting with b23712161d77e92842a8f593accfd2933e8c360d003454d4113d7d79c7d31bbd not found: ID does not exist" containerID="b23712161d77e92842a8f593accfd2933e8c360d003454d4113d7d79c7d31bbd" Mar 17 16:44:43 crc kubenswrapper[4767]: I0317 16:44:43.611420 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b23712161d77e92842a8f593accfd2933e8c360d003454d4113d7d79c7d31bbd"} err="failed to get container status \"b23712161d77e92842a8f593accfd2933e8c360d003454d4113d7d79c7d31bbd\": rpc error: code = NotFound desc = could not find container \"b23712161d77e92842a8f593accfd2933e8c360d003454d4113d7d79c7d31bbd\": container with ID starting with b23712161d77e92842a8f593accfd2933e8c360d003454d4113d7d79c7d31bbd not found: ID does not exist" Mar 17 16:44:45 crc kubenswrapper[4767]: I0317 16:44:45.375018 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3900db7-d204-4f6b-a0f3-79d172d16155" path="/var/lib/kubelet/pods/c3900db7-d204-4f6b-a0f3-79d172d16155/volumes" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.160909 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c"] Mar 17 16:45:00 crc kubenswrapper[4767]: E0317 16:45:00.162301 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerName="registry-server" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.162323 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerName="registry-server" Mar 17 16:45:00 crc kubenswrapper[4767]: E0317 16:45:00.162341 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerName="extract-utilities" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.162348 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerName="extract-utilities" Mar 17 16:45:00 crc kubenswrapper[4767]: E0317 16:45:00.162409 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerName="extract-content" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.162416 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerName="extract-content" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.162704 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3900db7-d204-4f6b-a0f3-79d172d16155" containerName="registry-server" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.163766 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.167063 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.168333 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.173193 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c"] Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.187941 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6b164e1-6caf-4305-af64-90f5c08ab830-config-volume\") pod \"collect-profiles-29562765-tm94c\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.188291 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6b164e1-6caf-4305-af64-90f5c08ab830-secret-volume\") pod \"collect-profiles-29562765-tm94c\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.188346 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c665t\" (UniqueName: \"kubernetes.io/projected/e6b164e1-6caf-4305-af64-90f5c08ab830-kube-api-access-c665t\") pod \"collect-profiles-29562765-tm94c\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.291178 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6b164e1-6caf-4305-af64-90f5c08ab830-secret-volume\") pod \"collect-profiles-29562765-tm94c\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.291260 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c665t\" (UniqueName: \"kubernetes.io/projected/e6b164e1-6caf-4305-af64-90f5c08ab830-kube-api-access-c665t\") pod \"collect-profiles-29562765-tm94c\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.291329 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6b164e1-6caf-4305-af64-90f5c08ab830-config-volume\") pod \"collect-profiles-29562765-tm94c\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.292254 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6b164e1-6caf-4305-af64-90f5c08ab830-config-volume\") pod \"collect-profiles-29562765-tm94c\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.300696 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6b164e1-6caf-4305-af64-90f5c08ab830-secret-volume\") pod \"collect-profiles-29562765-tm94c\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.316627 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c665t\" (UniqueName: \"kubernetes.io/projected/e6b164e1-6caf-4305-af64-90f5c08ab830-kube-api-access-c665t\") pod \"collect-profiles-29562765-tm94c\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:00 crc kubenswrapper[4767]: I0317 16:45:00.494989 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:01 crc kubenswrapper[4767]: I0317 16:45:01.012610 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c"] Mar 17 16:45:01 crc kubenswrapper[4767]: I0317 16:45:01.697504 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" event={"ID":"e6b164e1-6caf-4305-af64-90f5c08ab830","Type":"ContainerStarted","Data":"793ce8e35a0e5ae5242b3b84026b51f38aa849522471345ba87b9ff9076014ce"} Mar 17 16:45:01 crc kubenswrapper[4767]: I0317 16:45:01.697864 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" event={"ID":"e6b164e1-6caf-4305-af64-90f5c08ab830","Type":"ContainerStarted","Data":"aae79477ca92b214c3931f2c7da1d11764e4ccf166539377d1334b59cc9ee260"} Mar 17 16:45:01 crc kubenswrapper[4767]: I0317 16:45:01.734549 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" podStartSLOduration=1.734525007 podStartE2EDuration="1.734525007s" podCreationTimestamp="2026-03-17 16:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 16:45:01.727567437 +0000 UTC m=+4093.140883504" watchObservedRunningTime="2026-03-17 16:45:01.734525007 +0000 UTC m=+4093.147841054" Mar 17 16:45:02 crc kubenswrapper[4767]: I0317 16:45:02.711805 4767 generic.go:334] "Generic (PLEG): container finished" podID="e6b164e1-6caf-4305-af64-90f5c08ab830" containerID="793ce8e35a0e5ae5242b3b84026b51f38aa849522471345ba87b9ff9076014ce" exitCode=0 Mar 17 16:45:02 crc kubenswrapper[4767]: I0317 16:45:02.711871 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" event={"ID":"e6b164e1-6caf-4305-af64-90f5c08ab830","Type":"ContainerDied","Data":"793ce8e35a0e5ae5242b3b84026b51f38aa849522471345ba87b9ff9076014ce"} Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.211826 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.224320 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6b164e1-6caf-4305-af64-90f5c08ab830-secret-volume\") pod \"e6b164e1-6caf-4305-af64-90f5c08ab830\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.224430 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6b164e1-6caf-4305-af64-90f5c08ab830-config-volume\") pod \"e6b164e1-6caf-4305-af64-90f5c08ab830\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.224575 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c665t\" (UniqueName: \"kubernetes.io/projected/e6b164e1-6caf-4305-af64-90f5c08ab830-kube-api-access-c665t\") pod \"e6b164e1-6caf-4305-af64-90f5c08ab830\" (UID: \"e6b164e1-6caf-4305-af64-90f5c08ab830\") " Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.225275 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6b164e1-6caf-4305-af64-90f5c08ab830-config-volume" (OuterVolumeSpecName: "config-volume") pod "e6b164e1-6caf-4305-af64-90f5c08ab830" (UID: "e6b164e1-6caf-4305-af64-90f5c08ab830"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.241593 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6b164e1-6caf-4305-af64-90f5c08ab830-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e6b164e1-6caf-4305-af64-90f5c08ab830" (UID: "e6b164e1-6caf-4305-af64-90f5c08ab830"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.275742 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6b164e1-6caf-4305-af64-90f5c08ab830-kube-api-access-c665t" (OuterVolumeSpecName: "kube-api-access-c665t") pod "e6b164e1-6caf-4305-af64-90f5c08ab830" (UID: "e6b164e1-6caf-4305-af64-90f5c08ab830"). InnerVolumeSpecName "kube-api-access-c665t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.328981 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6b164e1-6caf-4305-af64-90f5c08ab830-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.329045 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6b164e1-6caf-4305-af64-90f5c08ab830-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.329062 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c665t\" (UniqueName: \"kubernetes.io/projected/e6b164e1-6caf-4305-af64-90f5c08ab830-kube-api-access-c665t\") on node \"crc\" DevicePath \"\"" Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.736018 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.735931 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562765-tm94c" event={"ID":"e6b164e1-6caf-4305-af64-90f5c08ab830","Type":"ContainerDied","Data":"aae79477ca92b214c3931f2c7da1d11764e4ccf166539377d1334b59cc9ee260"} Mar 17 16:45:04 crc kubenswrapper[4767]: I0317 16:45:04.739384 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aae79477ca92b214c3931f2c7da1d11764e4ccf166539377d1334b59cc9ee260" Mar 17 16:45:05 crc kubenswrapper[4767]: I0317 16:45:05.345125 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82"] Mar 17 16:45:05 crc kubenswrapper[4767]: I0317 16:45:05.376985 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562720-54h82"] Mar 17 16:45:07 crc kubenswrapper[4767]: I0317 16:45:07.373139 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62b3a4d3-4081-44b9-8d4b-fcebe2f89762" path="/var/lib/kubelet/pods/62b3a4d3-4081-44b9-8d4b-fcebe2f89762/volumes" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.332128 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8wvhf"] Mar 17 16:45:10 crc kubenswrapper[4767]: E0317 16:45:10.333035 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6b164e1-6caf-4305-af64-90f5c08ab830" containerName="collect-profiles" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.333051 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6b164e1-6caf-4305-af64-90f5c08ab830" containerName="collect-profiles" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.333350 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6b164e1-6caf-4305-af64-90f5c08ab830" containerName="collect-profiles" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.335393 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.349131 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8wvhf"] Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.380650 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-utilities\") pod \"community-operators-8wvhf\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.380731 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g898s\" (UniqueName: \"kubernetes.io/projected/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-kube-api-access-g898s\") pod \"community-operators-8wvhf\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.380762 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-catalog-content\") pod \"community-operators-8wvhf\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.484226 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-utilities\") pod \"community-operators-8wvhf\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.484303 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g898s\" (UniqueName: \"kubernetes.io/projected/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-kube-api-access-g898s\") pod \"community-operators-8wvhf\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.484351 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-catalog-content\") pod \"community-operators-8wvhf\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.485147 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-catalog-content\") pod \"community-operators-8wvhf\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.485457 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-utilities\") pod \"community-operators-8wvhf\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.515710 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g898s\" (UniqueName: \"kubernetes.io/projected/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-kube-api-access-g898s\") pod \"community-operators-8wvhf\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.535572 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b7hs4"] Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.539596 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.565036 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7hs4"] Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.589167 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88ghl\" (UniqueName: \"kubernetes.io/projected/967a6ff6-aab8-4944-bd46-49034d1ff5b7-kube-api-access-88ghl\") pod \"redhat-marketplace-b7hs4\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.589320 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-utilities\") pod \"redhat-marketplace-b7hs4\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.589378 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-catalog-content\") pod \"redhat-marketplace-b7hs4\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.657428 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.692556 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-utilities\") pod \"redhat-marketplace-b7hs4\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.692643 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-catalog-content\") pod \"redhat-marketplace-b7hs4\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.693260 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-utilities\") pod \"redhat-marketplace-b7hs4\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.693817 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88ghl\" (UniqueName: \"kubernetes.io/projected/967a6ff6-aab8-4944-bd46-49034d1ff5b7-kube-api-access-88ghl\") pod \"redhat-marketplace-b7hs4\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.693270 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-catalog-content\") pod \"redhat-marketplace-b7hs4\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.720084 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88ghl\" (UniqueName: \"kubernetes.io/projected/967a6ff6-aab8-4944-bd46-49034d1ff5b7-kube-api-access-88ghl\") pod \"redhat-marketplace-b7hs4\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:10 crc kubenswrapper[4767]: I0317 16:45:10.906783 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:11 crc kubenswrapper[4767]: I0317 16:45:11.338224 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8wvhf"] Mar 17 16:45:11 crc kubenswrapper[4767]: I0317 16:45:11.609098 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7hs4"] Mar 17 16:45:11 crc kubenswrapper[4767]: W0317 16:45:11.612992 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod967a6ff6_aab8_4944_bd46_49034d1ff5b7.slice/crio-a52192c2e8d65787547cf7cc536af4676a2ea0a05e49056d51913a633e3a722b WatchSource:0}: Error finding container a52192c2e8d65787547cf7cc536af4676a2ea0a05e49056d51913a633e3a722b: Status 404 returned error can't find the container with id a52192c2e8d65787547cf7cc536af4676a2ea0a05e49056d51913a633e3a722b Mar 17 16:45:12 crc kubenswrapper[4767]: I0317 16:45:12.052538 4767 generic.go:334] "Generic (PLEG): container finished" podID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" containerID="e3d0693dd4dc9b84894e6dbb249176c0708920ffe83f0783bf734bedeaf24c12" exitCode=0 Mar 17 16:45:12 crc kubenswrapper[4767]: I0317 16:45:12.052761 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7hs4" event={"ID":"967a6ff6-aab8-4944-bd46-49034d1ff5b7","Type":"ContainerDied","Data":"e3d0693dd4dc9b84894e6dbb249176c0708920ffe83f0783bf734bedeaf24c12"} Mar 17 16:45:12 crc kubenswrapper[4767]: I0317 16:45:12.052921 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7hs4" event={"ID":"967a6ff6-aab8-4944-bd46-49034d1ff5b7","Type":"ContainerStarted","Data":"a52192c2e8d65787547cf7cc536af4676a2ea0a05e49056d51913a633e3a722b"} Mar 17 16:45:12 crc kubenswrapper[4767]: I0317 16:45:12.055092 4767 generic.go:334] "Generic (PLEG): container finished" podID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" containerID="7e4aeb90bd648d2a388690bfbbdabdd40eaa35e26ff9f063f5781a397993d650" exitCode=0 Mar 17 16:45:12 crc kubenswrapper[4767]: I0317 16:45:12.055134 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wvhf" event={"ID":"7ac92cf7-2cc9-4f41-a144-b5908c4279cd","Type":"ContainerDied","Data":"7e4aeb90bd648d2a388690bfbbdabdd40eaa35e26ff9f063f5781a397993d650"} Mar 17 16:45:12 crc kubenswrapper[4767]: I0317 16:45:12.055162 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wvhf" event={"ID":"7ac92cf7-2cc9-4f41-a144-b5908c4279cd","Type":"ContainerStarted","Data":"66977789c6db5ec79babd93a17737da687c844440fa1b78112f58ed97ee92ae6"} Mar 17 16:45:12 crc kubenswrapper[4767]: I0317 16:45:12.055927 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 16:45:14 crc kubenswrapper[4767]: I0317 16:45:14.083159 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7hs4" event={"ID":"967a6ff6-aab8-4944-bd46-49034d1ff5b7","Type":"ContainerStarted","Data":"5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273"} Mar 17 16:45:14 crc kubenswrapper[4767]: I0317 16:45:14.087598 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wvhf" event={"ID":"7ac92cf7-2cc9-4f41-a144-b5908c4279cd","Type":"ContainerStarted","Data":"f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41"} Mar 17 16:45:15 crc kubenswrapper[4767]: I0317 16:45:15.103241 4767 generic.go:334] "Generic (PLEG): container finished" podID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" containerID="5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273" exitCode=0 Mar 17 16:45:15 crc kubenswrapper[4767]: I0317 16:45:15.103340 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7hs4" event={"ID":"967a6ff6-aab8-4944-bd46-49034d1ff5b7","Type":"ContainerDied","Data":"5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273"} Mar 17 16:45:15 crc kubenswrapper[4767]: I0317 16:45:15.107016 4767 generic.go:334] "Generic (PLEG): container finished" podID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" containerID="f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41" exitCode=0 Mar 17 16:45:15 crc kubenswrapper[4767]: I0317 16:45:15.107068 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wvhf" event={"ID":"7ac92cf7-2cc9-4f41-a144-b5908c4279cd","Type":"ContainerDied","Data":"f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41"} Mar 17 16:45:16 crc kubenswrapper[4767]: I0317 16:45:16.122478 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7hs4" event={"ID":"967a6ff6-aab8-4944-bd46-49034d1ff5b7","Type":"ContainerStarted","Data":"1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb"} Mar 17 16:45:16 crc kubenswrapper[4767]: I0317 16:45:16.128394 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wvhf" event={"ID":"7ac92cf7-2cc9-4f41-a144-b5908c4279cd","Type":"ContainerStarted","Data":"c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0"} Mar 17 16:45:16 crc kubenswrapper[4767]: I0317 16:45:16.195200 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b7hs4" podStartSLOduration=2.692749734 podStartE2EDuration="6.195154071s" podCreationTimestamp="2026-03-17 16:45:10 +0000 UTC" firstStartedPulling="2026-03-17 16:45:12.055685305 +0000 UTC m=+4103.469001352" lastFinishedPulling="2026-03-17 16:45:15.558089632 +0000 UTC m=+4106.971405689" observedRunningTime="2026-03-17 16:45:16.143612988 +0000 UTC m=+4107.556929045" watchObservedRunningTime="2026-03-17 16:45:16.195154071 +0000 UTC m=+4107.608470118" Mar 17 16:45:16 crc kubenswrapper[4767]: I0317 16:45:16.211015 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8wvhf" podStartSLOduration=2.743486804 podStartE2EDuration="6.210983336s" podCreationTimestamp="2026-03-17 16:45:10 +0000 UTC" firstStartedPulling="2026-03-17 16:45:12.056997563 +0000 UTC m=+4103.470313600" lastFinishedPulling="2026-03-17 16:45:15.524494085 +0000 UTC m=+4106.937810132" observedRunningTime="2026-03-17 16:45:16.183253028 +0000 UTC m=+4107.596569095" watchObservedRunningTime="2026-03-17 16:45:16.210983336 +0000 UTC m=+4107.624299383" Mar 17 16:45:20 crc kubenswrapper[4767]: I0317 16:45:20.659303 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:20 crc kubenswrapper[4767]: I0317 16:45:20.659932 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:20 crc kubenswrapper[4767]: I0317 16:45:20.731044 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:20 crc kubenswrapper[4767]: I0317 16:45:20.908638 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:20 crc kubenswrapper[4767]: I0317 16:45:20.909015 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:20 crc kubenswrapper[4767]: I0317 16:45:20.968496 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:21 crc kubenswrapper[4767]: I0317 16:45:21.240659 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:21 crc kubenswrapper[4767]: I0317 16:45:21.248584 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:22 crc kubenswrapper[4767]: I0317 16:45:22.907644 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7hs4"] Mar 17 16:45:23 crc kubenswrapper[4767]: I0317 16:45:23.207156 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b7hs4" podUID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" containerName="registry-server" containerID="cri-o://1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb" gracePeriod=2 Mar 17 16:45:23 crc kubenswrapper[4767]: I0317 16:45:23.256268 4767 scope.go:117] "RemoveContainer" containerID="27ad7e5eaab2a567614bd41747169bac163317f973daeb5fdbdce15ecd3695e6" Mar 17 16:45:23 crc kubenswrapper[4767]: I0317 16:45:23.522903 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8wvhf"] Mar 17 16:45:23 crc kubenswrapper[4767]: I0317 16:45:23.523200 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8wvhf" podUID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" containerName="registry-server" containerID="cri-o://c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0" gracePeriod=2 Mar 17 16:45:23 crc kubenswrapper[4767]: I0317 16:45:23.795048 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:23 crc kubenswrapper[4767]: I0317 16:45:23.929332 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88ghl\" (UniqueName: \"kubernetes.io/projected/967a6ff6-aab8-4944-bd46-49034d1ff5b7-kube-api-access-88ghl\") pod \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " Mar 17 16:45:23 crc kubenswrapper[4767]: I0317 16:45:23.929576 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-utilities\") pod \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " Mar 17 16:45:23 crc kubenswrapper[4767]: I0317 16:45:23.929796 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-catalog-content\") pod \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\" (UID: \"967a6ff6-aab8-4944-bd46-49034d1ff5b7\") " Mar 17 16:45:23 crc kubenswrapper[4767]: I0317 16:45:23.930907 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-utilities" (OuterVolumeSpecName: "utilities") pod "967a6ff6-aab8-4944-bd46-49034d1ff5b7" (UID: "967a6ff6-aab8-4944-bd46-49034d1ff5b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:45:23 crc kubenswrapper[4767]: I0317 16:45:23.937523 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/967a6ff6-aab8-4944-bd46-49034d1ff5b7-kube-api-access-88ghl" (OuterVolumeSpecName: "kube-api-access-88ghl") pod "967a6ff6-aab8-4944-bd46-49034d1ff5b7" (UID: "967a6ff6-aab8-4944-bd46-49034d1ff5b7"). InnerVolumeSpecName "kube-api-access-88ghl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:45:23 crc kubenswrapper[4767]: I0317 16:45:23.965576 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "967a6ff6-aab8-4944-bd46-49034d1ff5b7" (UID: "967a6ff6-aab8-4944-bd46-49034d1ff5b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.034839 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.035351 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88ghl\" (UniqueName: \"kubernetes.io/projected/967a6ff6-aab8-4944-bd46-49034d1ff5b7-kube-api-access-88ghl\") on node \"crc\" DevicePath \"\"" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.035463 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.035554 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/967a6ff6-aab8-4944-bd46-49034d1ff5b7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.143582 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-utilities\") pod \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.144133 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-catalog-content\") pod \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.144483 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g898s\" (UniqueName: \"kubernetes.io/projected/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-kube-api-access-g898s\") pod \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\" (UID: \"7ac92cf7-2cc9-4f41-a144-b5908c4279cd\") " Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.146358 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-utilities" (OuterVolumeSpecName: "utilities") pod "7ac92cf7-2cc9-4f41-a144-b5908c4279cd" (UID: "7ac92cf7-2cc9-4f41-a144-b5908c4279cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.151719 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-kube-api-access-g898s" (OuterVolumeSpecName: "kube-api-access-g898s") pod "7ac92cf7-2cc9-4f41-a144-b5908c4279cd" (UID: "7ac92cf7-2cc9-4f41-a144-b5908c4279cd"). InnerVolumeSpecName "kube-api-access-g898s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.166702 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.239125 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ac92cf7-2cc9-4f41-a144-b5908c4279cd" (UID: "7ac92cf7-2cc9-4f41-a144-b5908c4279cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.247908 4767 generic.go:334] "Generic (PLEG): container finished" podID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" containerID="1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb" exitCode=0 Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.247981 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7hs4" event={"ID":"967a6ff6-aab8-4944-bd46-49034d1ff5b7","Type":"ContainerDied","Data":"1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb"} Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.248012 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7hs4" event={"ID":"967a6ff6-aab8-4944-bd46-49034d1ff5b7","Type":"ContainerDied","Data":"a52192c2e8d65787547cf7cc536af4676a2ea0a05e49056d51913a633e3a722b"} Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.248033 4767 scope.go:117] "RemoveContainer" containerID="1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.248233 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7hs4" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.254594 4767 generic.go:334] "Generic (PLEG): container finished" podID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" containerID="c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0" exitCode=0 Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.254662 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wvhf" event={"ID":"7ac92cf7-2cc9-4f41-a144-b5908c4279cd","Type":"ContainerDied","Data":"c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0"} Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.254671 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8wvhf" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.254700 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wvhf" event={"ID":"7ac92cf7-2cc9-4f41-a144-b5908c4279cd","Type":"ContainerDied","Data":"66977789c6db5ec79babd93a17737da687c844440fa1b78112f58ed97ee92ae6"} Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.268879 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.268916 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g898s\" (UniqueName: \"kubernetes.io/projected/7ac92cf7-2cc9-4f41-a144-b5908c4279cd-kube-api-access-g898s\") on node \"crc\" DevicePath \"\"" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.285689 4767 scope.go:117] "RemoveContainer" containerID="5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.305259 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7hs4"] Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.321325 4767 scope.go:117] "RemoveContainer" containerID="e3d0693dd4dc9b84894e6dbb249176c0708920ffe83f0783bf734bedeaf24c12" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.323674 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7hs4"] Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.334945 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8wvhf"] Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.346294 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8wvhf"] Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.397216 4767 scope.go:117] "RemoveContainer" containerID="1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb" Mar 17 16:45:24 crc kubenswrapper[4767]: E0317 16:45:24.397579 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb\": container with ID starting with 1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb not found: ID does not exist" containerID="1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.397623 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb"} err="failed to get container status \"1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb\": rpc error: code = NotFound desc = could not find container \"1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb\": container with ID starting with 1b0423c9763f3f77d9479e8fb383d79dcb6cadc0bf26d6c1c81d3d312357e3cb not found: ID does not exist" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.397649 4767 scope.go:117] "RemoveContainer" containerID="5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273" Mar 17 16:45:24 crc kubenswrapper[4767]: E0317 16:45:24.398152 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273\": container with ID starting with 5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273 not found: ID does not exist" containerID="5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.398358 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273"} err="failed to get container status \"5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273\": rpc error: code = NotFound desc = could not find container \"5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273\": container with ID starting with 5d388fbafd15be79b02ea08f4ac9038446c12f354b833a5cc92ac2975914e273 not found: ID does not exist" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.398377 4767 scope.go:117] "RemoveContainer" containerID="e3d0693dd4dc9b84894e6dbb249176c0708920ffe83f0783bf734bedeaf24c12" Mar 17 16:45:24 crc kubenswrapper[4767]: E0317 16:45:24.398748 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3d0693dd4dc9b84894e6dbb249176c0708920ffe83f0783bf734bedeaf24c12\": container with ID starting with e3d0693dd4dc9b84894e6dbb249176c0708920ffe83f0783bf734bedeaf24c12 not found: ID does not exist" containerID="e3d0693dd4dc9b84894e6dbb249176c0708920ffe83f0783bf734bedeaf24c12" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.398773 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3d0693dd4dc9b84894e6dbb249176c0708920ffe83f0783bf734bedeaf24c12"} err="failed to get container status \"e3d0693dd4dc9b84894e6dbb249176c0708920ffe83f0783bf734bedeaf24c12\": rpc error: code = NotFound desc = could not find container \"e3d0693dd4dc9b84894e6dbb249176c0708920ffe83f0783bf734bedeaf24c12\": container with ID starting with e3d0693dd4dc9b84894e6dbb249176c0708920ffe83f0783bf734bedeaf24c12 not found: ID does not exist" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.398787 4767 scope.go:117] "RemoveContainer" containerID="c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.459713 4767 scope.go:117] "RemoveContainer" containerID="f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.490044 4767 scope.go:117] "RemoveContainer" containerID="7e4aeb90bd648d2a388690bfbbdabdd40eaa35e26ff9f063f5781a397993d650" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.553527 4767 scope.go:117] "RemoveContainer" containerID="c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0" Mar 17 16:45:24 crc kubenswrapper[4767]: E0317 16:45:24.554061 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0\": container with ID starting with c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0 not found: ID does not exist" containerID="c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.554189 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0"} err="failed to get container status \"c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0\": rpc error: code = NotFound desc = could not find container \"c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0\": container with ID starting with c73a9f9fc605d63bf67cef6b7333144ee486a2d2ad8170204bb2e3712fab2dc0 not found: ID does not exist" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.554294 4767 scope.go:117] "RemoveContainer" containerID="f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41" Mar 17 16:45:24 crc kubenswrapper[4767]: E0317 16:45:24.554691 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41\": container with ID starting with f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41 not found: ID does not exist" containerID="f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.554733 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41"} err="failed to get container status \"f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41\": rpc error: code = NotFound desc = could not find container \"f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41\": container with ID starting with f44121557624ebebfbc94cbace55fb3a5310944f1dc65100f0356a0d53a54a41 not found: ID does not exist" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.554764 4767 scope.go:117] "RemoveContainer" containerID="7e4aeb90bd648d2a388690bfbbdabdd40eaa35e26ff9f063f5781a397993d650" Mar 17 16:45:24 crc kubenswrapper[4767]: E0317 16:45:24.555015 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e4aeb90bd648d2a388690bfbbdabdd40eaa35e26ff9f063f5781a397993d650\": container with ID starting with 7e4aeb90bd648d2a388690bfbbdabdd40eaa35e26ff9f063f5781a397993d650 not found: ID does not exist" containerID="7e4aeb90bd648d2a388690bfbbdabdd40eaa35e26ff9f063f5781a397993d650" Mar 17 16:45:24 crc kubenswrapper[4767]: I0317 16:45:24.555043 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e4aeb90bd648d2a388690bfbbdabdd40eaa35e26ff9f063f5781a397993d650"} err="failed to get container status \"7e4aeb90bd648d2a388690bfbbdabdd40eaa35e26ff9f063f5781a397993d650\": rpc error: code = NotFound desc = could not find container \"7e4aeb90bd648d2a388690bfbbdabdd40eaa35e26ff9f063f5781a397993d650\": container with ID starting with 7e4aeb90bd648d2a388690bfbbdabdd40eaa35e26ff9f063f5781a397993d650 not found: ID does not exist" Mar 17 16:45:25 crc kubenswrapper[4767]: I0317 16:45:25.373345 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" path="/var/lib/kubelet/pods/7ac92cf7-2cc9-4f41-a144-b5908c4279cd/volumes" Mar 17 16:45:25 crc kubenswrapper[4767]: I0317 16:45:25.376484 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" path="/var/lib/kubelet/pods/967a6ff6-aab8-4944-bd46-49034d1ff5b7/volumes" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.165630 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562766-wrpqm"] Mar 17 16:46:00 crc kubenswrapper[4767]: E0317 16:46:00.167263 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" containerName="extract-content" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.167285 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" containerName="extract-content" Mar 17 16:46:00 crc kubenswrapper[4767]: E0317 16:46:00.167298 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" containerName="registry-server" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.167305 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" containerName="registry-server" Mar 17 16:46:00 crc kubenswrapper[4767]: E0317 16:46:00.167331 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" containerName="extract-content" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.167340 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" containerName="extract-content" Mar 17 16:46:00 crc kubenswrapper[4767]: E0317 16:46:00.167365 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" containerName="extract-utilities" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.167374 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" containerName="extract-utilities" Mar 17 16:46:00 crc kubenswrapper[4767]: E0317 16:46:00.167394 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" containerName="registry-server" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.167401 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" containerName="registry-server" Mar 17 16:46:00 crc kubenswrapper[4767]: E0317 16:46:00.167432 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" containerName="extract-utilities" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.167440 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" containerName="extract-utilities" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.167796 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="967a6ff6-aab8-4944-bd46-49034d1ff5b7" containerName="registry-server" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.167847 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac92cf7-2cc9-4f41-a144-b5908c4279cd" containerName="registry-server" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.169240 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562766-wrpqm" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.172520 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.172769 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.172980 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.179147 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562766-wrpqm"] Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.341816 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5cwn\" (UniqueName: \"kubernetes.io/projected/933c4014-a9df-4fca-a5fb-ce61f79403c8-kube-api-access-l5cwn\") pod \"auto-csr-approver-29562766-wrpqm\" (UID: \"933c4014-a9df-4fca-a5fb-ce61f79403c8\") " pod="openshift-infra/auto-csr-approver-29562766-wrpqm" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.444480 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5cwn\" (UniqueName: \"kubernetes.io/projected/933c4014-a9df-4fca-a5fb-ce61f79403c8-kube-api-access-l5cwn\") pod \"auto-csr-approver-29562766-wrpqm\" (UID: \"933c4014-a9df-4fca-a5fb-ce61f79403c8\") " pod="openshift-infra/auto-csr-approver-29562766-wrpqm" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.468410 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5cwn\" (UniqueName: \"kubernetes.io/projected/933c4014-a9df-4fca-a5fb-ce61f79403c8-kube-api-access-l5cwn\") pod \"auto-csr-approver-29562766-wrpqm\" (UID: \"933c4014-a9df-4fca-a5fb-ce61f79403c8\") " pod="openshift-infra/auto-csr-approver-29562766-wrpqm" Mar 17 16:46:00 crc kubenswrapper[4767]: I0317 16:46:00.514634 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562766-wrpqm" Mar 17 16:46:01 crc kubenswrapper[4767]: I0317 16:46:01.014740 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562766-wrpqm"] Mar 17 16:46:01 crc kubenswrapper[4767]: I0317 16:46:01.715547 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562766-wrpqm" event={"ID":"933c4014-a9df-4fca-a5fb-ce61f79403c8","Type":"ContainerStarted","Data":"1051d6a1834446bf3dfbddce57afe0929cb2d217ee0424b9f7455661b715ed70"} Mar 17 16:46:03 crc kubenswrapper[4767]: I0317 16:46:03.745145 4767 generic.go:334] "Generic (PLEG): container finished" podID="933c4014-a9df-4fca-a5fb-ce61f79403c8" containerID="d033154a465879b6aa92090e93c2d181b85605cfe09cc5cecfeb8dd0043da031" exitCode=0 Mar 17 16:46:03 crc kubenswrapper[4767]: I0317 16:46:03.745462 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562766-wrpqm" event={"ID":"933c4014-a9df-4fca-a5fb-ce61f79403c8","Type":"ContainerDied","Data":"d033154a465879b6aa92090e93c2d181b85605cfe09cc5cecfeb8dd0043da031"} Mar 17 16:46:05 crc kubenswrapper[4767]: I0317 16:46:05.190991 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562766-wrpqm" Mar 17 16:46:05 crc kubenswrapper[4767]: I0317 16:46:05.289482 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5cwn\" (UniqueName: \"kubernetes.io/projected/933c4014-a9df-4fca-a5fb-ce61f79403c8-kube-api-access-l5cwn\") pod \"933c4014-a9df-4fca-a5fb-ce61f79403c8\" (UID: \"933c4014-a9df-4fca-a5fb-ce61f79403c8\") " Mar 17 16:46:05 crc kubenswrapper[4767]: I0317 16:46:05.302338 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/933c4014-a9df-4fca-a5fb-ce61f79403c8-kube-api-access-l5cwn" (OuterVolumeSpecName: "kube-api-access-l5cwn") pod "933c4014-a9df-4fca-a5fb-ce61f79403c8" (UID: "933c4014-a9df-4fca-a5fb-ce61f79403c8"). InnerVolumeSpecName "kube-api-access-l5cwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:46:05 crc kubenswrapper[4767]: I0317 16:46:05.392561 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5cwn\" (UniqueName: \"kubernetes.io/projected/933c4014-a9df-4fca-a5fb-ce61f79403c8-kube-api-access-l5cwn\") on node \"crc\" DevicePath \"\"" Mar 17 16:46:05 crc kubenswrapper[4767]: I0317 16:46:05.774499 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562766-wrpqm" event={"ID":"933c4014-a9df-4fca-a5fb-ce61f79403c8","Type":"ContainerDied","Data":"1051d6a1834446bf3dfbddce57afe0929cb2d217ee0424b9f7455661b715ed70"} Mar 17 16:46:05 crc kubenswrapper[4767]: I0317 16:46:05.774572 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1051d6a1834446bf3dfbddce57afe0929cb2d217ee0424b9f7455661b715ed70" Mar 17 16:46:05 crc kubenswrapper[4767]: I0317 16:46:05.774670 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562766-wrpqm" Mar 17 16:46:06 crc kubenswrapper[4767]: I0317 16:46:06.273045 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562760-k2rwn"] Mar 17 16:46:06 crc kubenswrapper[4767]: I0317 16:46:06.285645 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562760-k2rwn"] Mar 17 16:46:07 crc kubenswrapper[4767]: I0317 16:46:07.378925 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b41df7a-d8a2-4883-8018-f306322bc3da" path="/var/lib/kubelet/pods/4b41df7a-d8a2-4883-8018-f306322bc3da/volumes" Mar 17 16:46:23 crc kubenswrapper[4767]: I0317 16:46:23.642537 4767 scope.go:117] "RemoveContainer" containerID="9c4929e311dc15005520721e20da67c4f563c8416bc13c29c5104776f62d5ea8" Mar 17 16:46:34 crc kubenswrapper[4767]: I0317 16:46:34.166948 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:46:34 crc kubenswrapper[4767]: I0317 16:46:34.167921 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:47:04 crc kubenswrapper[4767]: I0317 16:47:04.166907 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:47:04 crc kubenswrapper[4767]: I0317 16:47:04.168868 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:47:34 crc kubenswrapper[4767]: I0317 16:47:34.166445 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:47:34 crc kubenswrapper[4767]: I0317 16:47:34.167960 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:47:34 crc kubenswrapper[4767]: I0317 16:47:34.168124 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 16:47:34 crc kubenswrapper[4767]: I0317 16:47:34.169539 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a329b73e6c74a32ee0eed732b0cfe2e5edcb9ecafe799acd643961d80496c7af"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 16:47:34 crc kubenswrapper[4767]: I0317 16:47:34.169724 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://a329b73e6c74a32ee0eed732b0cfe2e5edcb9ecafe799acd643961d80496c7af" gracePeriod=600 Mar 17 16:47:35 crc kubenswrapper[4767]: I0317 16:47:35.093466 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="a329b73e6c74a32ee0eed732b0cfe2e5edcb9ecafe799acd643961d80496c7af" exitCode=0 Mar 17 16:47:35 crc kubenswrapper[4767]: I0317 16:47:35.093603 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"a329b73e6c74a32ee0eed732b0cfe2e5edcb9ecafe799acd643961d80496c7af"} Mar 17 16:47:35 crc kubenswrapper[4767]: I0317 16:47:35.094289 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216"} Mar 17 16:47:35 crc kubenswrapper[4767]: I0317 16:47:35.094326 4767 scope.go:117] "RemoveContainer" containerID="cc18665cf910f2acb57f785103363155cd2316c2f45848732a3450a8e7457055" Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.155993 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562768-2prsp"] Mar 17 16:48:00 crc kubenswrapper[4767]: E0317 16:48:00.157202 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933c4014-a9df-4fca-a5fb-ce61f79403c8" containerName="oc" Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.157223 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="933c4014-a9df-4fca-a5fb-ce61f79403c8" containerName="oc" Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.157588 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="933c4014-a9df-4fca-a5fb-ce61f79403c8" containerName="oc" Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.158657 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562768-2prsp" Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.165708 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.165774 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.166158 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.168339 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562768-2prsp"] Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.196765 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzvd6\" (UniqueName: \"kubernetes.io/projected/7f471d2c-947d-455c-8a41-077987d2d293-kube-api-access-kzvd6\") pod \"auto-csr-approver-29562768-2prsp\" (UID: \"7f471d2c-947d-455c-8a41-077987d2d293\") " pod="openshift-infra/auto-csr-approver-29562768-2prsp" Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.300778 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzvd6\" (UniqueName: \"kubernetes.io/projected/7f471d2c-947d-455c-8a41-077987d2d293-kube-api-access-kzvd6\") pod \"auto-csr-approver-29562768-2prsp\" (UID: \"7f471d2c-947d-455c-8a41-077987d2d293\") " pod="openshift-infra/auto-csr-approver-29562768-2prsp" Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.332586 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzvd6\" (UniqueName: \"kubernetes.io/projected/7f471d2c-947d-455c-8a41-077987d2d293-kube-api-access-kzvd6\") pod \"auto-csr-approver-29562768-2prsp\" (UID: \"7f471d2c-947d-455c-8a41-077987d2d293\") " pod="openshift-infra/auto-csr-approver-29562768-2prsp" Mar 17 16:48:00 crc kubenswrapper[4767]: I0317 16:48:00.501731 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562768-2prsp" Mar 17 16:48:01 crc kubenswrapper[4767]: I0317 16:48:01.025248 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562768-2prsp"] Mar 17 16:48:01 crc kubenswrapper[4767]: I0317 16:48:01.404140 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562768-2prsp" event={"ID":"7f471d2c-947d-455c-8a41-077987d2d293","Type":"ContainerStarted","Data":"a448834aada28b290eb8ff812a0943cc957ebdfd17ebbb38565d3e9a2e63165e"} Mar 17 16:48:03 crc kubenswrapper[4767]: I0317 16:48:03.449720 4767 generic.go:334] "Generic (PLEG): container finished" podID="7f471d2c-947d-455c-8a41-077987d2d293" containerID="5d11702509f4ab6589b4b0bcdb90f37497ecd810d611c6652bfb9a5820ae38ff" exitCode=0 Mar 17 16:48:03 crc kubenswrapper[4767]: I0317 16:48:03.449796 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562768-2prsp" event={"ID":"7f471d2c-947d-455c-8a41-077987d2d293","Type":"ContainerDied","Data":"5d11702509f4ab6589b4b0bcdb90f37497ecd810d611c6652bfb9a5820ae38ff"} Mar 17 16:48:04 crc kubenswrapper[4767]: I0317 16:48:04.915313 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562768-2prsp" Mar 17 16:48:05 crc kubenswrapper[4767]: I0317 16:48:05.092053 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzvd6\" (UniqueName: \"kubernetes.io/projected/7f471d2c-947d-455c-8a41-077987d2d293-kube-api-access-kzvd6\") pod \"7f471d2c-947d-455c-8a41-077987d2d293\" (UID: \"7f471d2c-947d-455c-8a41-077987d2d293\") " Mar 17 16:48:05 crc kubenswrapper[4767]: I0317 16:48:05.099609 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f471d2c-947d-455c-8a41-077987d2d293-kube-api-access-kzvd6" (OuterVolumeSpecName: "kube-api-access-kzvd6") pod "7f471d2c-947d-455c-8a41-077987d2d293" (UID: "7f471d2c-947d-455c-8a41-077987d2d293"). InnerVolumeSpecName "kube-api-access-kzvd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:48:05 crc kubenswrapper[4767]: I0317 16:48:05.196289 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzvd6\" (UniqueName: \"kubernetes.io/projected/7f471d2c-947d-455c-8a41-077987d2d293-kube-api-access-kzvd6\") on node \"crc\" DevicePath \"\"" Mar 17 16:48:05 crc kubenswrapper[4767]: I0317 16:48:05.477132 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562768-2prsp" event={"ID":"7f471d2c-947d-455c-8a41-077987d2d293","Type":"ContainerDied","Data":"a448834aada28b290eb8ff812a0943cc957ebdfd17ebbb38565d3e9a2e63165e"} Mar 17 16:48:05 crc kubenswrapper[4767]: I0317 16:48:05.477191 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a448834aada28b290eb8ff812a0943cc957ebdfd17ebbb38565d3e9a2e63165e" Mar 17 16:48:05 crc kubenswrapper[4767]: I0317 16:48:05.477258 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562768-2prsp" Mar 17 16:48:06 crc kubenswrapper[4767]: I0317 16:48:06.000599 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562762-qmqmf"] Mar 17 16:48:06 crc kubenswrapper[4767]: I0317 16:48:06.013087 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562762-qmqmf"] Mar 17 16:48:07 crc kubenswrapper[4767]: I0317 16:48:07.663808 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d767c586-6134-4c00-ab0e-dd16cf2934fa" path="/var/lib/kubelet/pods/d767c586-6134-4c00-ab0e-dd16cf2934fa/volumes" Mar 17 16:48:23 crc kubenswrapper[4767]: I0317 16:48:23.858274 4767 scope.go:117] "RemoveContainer" containerID="e7ff1ec459e8edbc1902c2fb66b89ddeb8e022e23ed5e41840950a15c5405291" Mar 17 16:49:34 crc kubenswrapper[4767]: I0317 16:49:34.166629 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:49:34 crc kubenswrapper[4767]: I0317 16:49:34.167268 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.163165 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562770-njfq7"] Mar 17 16:50:00 crc kubenswrapper[4767]: E0317 16:50:00.164599 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f471d2c-947d-455c-8a41-077987d2d293" containerName="oc" Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.164634 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f471d2c-947d-455c-8a41-077987d2d293" containerName="oc" Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.164975 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f471d2c-947d-455c-8a41-077987d2d293" containerName="oc" Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.167860 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562770-njfq7" Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.175714 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.175999 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.176591 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.184466 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562770-njfq7"] Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.243149 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5wqj\" (UniqueName: \"kubernetes.io/projected/64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2-kube-api-access-h5wqj\") pod \"auto-csr-approver-29562770-njfq7\" (UID: \"64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2\") " pod="openshift-infra/auto-csr-approver-29562770-njfq7" Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.346942 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5wqj\" (UniqueName: \"kubernetes.io/projected/64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2-kube-api-access-h5wqj\") pod \"auto-csr-approver-29562770-njfq7\" (UID: \"64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2\") " pod="openshift-infra/auto-csr-approver-29562770-njfq7" Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.375210 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5wqj\" (UniqueName: \"kubernetes.io/projected/64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2-kube-api-access-h5wqj\") pod \"auto-csr-approver-29562770-njfq7\" (UID: \"64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2\") " pod="openshift-infra/auto-csr-approver-29562770-njfq7" Mar 17 16:50:00 crc kubenswrapper[4767]: I0317 16:50:00.493131 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562770-njfq7" Mar 17 16:50:01 crc kubenswrapper[4767]: I0317 16:50:01.011933 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562770-njfq7"] Mar 17 16:50:01 crc kubenswrapper[4767]: I0317 16:50:01.823232 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562770-njfq7" event={"ID":"64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2","Type":"ContainerStarted","Data":"1d515a8aea2cd038cb60c8bd0f4963f63d7cf6cfc6d24d1737965d6c67dc0393"} Mar 17 16:50:02 crc kubenswrapper[4767]: I0317 16:50:02.837689 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562770-njfq7" event={"ID":"64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2","Type":"ContainerStarted","Data":"64888c8523b4e039327ffdddd6972f0fdb2ecd8400989b45bd137355fc3ce45d"} Mar 17 16:50:02 crc kubenswrapper[4767]: I0317 16:50:02.864304 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562770-njfq7" podStartSLOduration=1.5858334969999999 podStartE2EDuration="2.864266463s" podCreationTimestamp="2026-03-17 16:50:00 +0000 UTC" firstStartedPulling="2026-03-17 16:50:01.011948784 +0000 UTC m=+4392.425264831" lastFinishedPulling="2026-03-17 16:50:02.29038176 +0000 UTC m=+4393.703697797" observedRunningTime="2026-03-17 16:50:02.854806755 +0000 UTC m=+4394.268122822" watchObservedRunningTime="2026-03-17 16:50:02.864266463 +0000 UTC m=+4394.277582520" Mar 17 16:50:03 crc kubenswrapper[4767]: I0317 16:50:03.854280 4767 generic.go:334] "Generic (PLEG): container finished" podID="64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2" containerID="64888c8523b4e039327ffdddd6972f0fdb2ecd8400989b45bd137355fc3ce45d" exitCode=0 Mar 17 16:50:03 crc kubenswrapper[4767]: I0317 16:50:03.855643 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562770-njfq7" event={"ID":"64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2","Type":"ContainerDied","Data":"64888c8523b4e039327ffdddd6972f0fdb2ecd8400989b45bd137355fc3ce45d"} Mar 17 16:50:04 crc kubenswrapper[4767]: I0317 16:50:04.166395 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:50:04 crc kubenswrapper[4767]: I0317 16:50:04.166521 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:50:05 crc kubenswrapper[4767]: I0317 16:50:05.296711 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562770-njfq7" Mar 17 16:50:05 crc kubenswrapper[4767]: I0317 16:50:05.335773 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5wqj\" (UniqueName: \"kubernetes.io/projected/64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2-kube-api-access-h5wqj\") pod \"64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2\" (UID: \"64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2\") " Mar 17 16:50:05 crc kubenswrapper[4767]: I0317 16:50:05.345872 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2-kube-api-access-h5wqj" (OuterVolumeSpecName: "kube-api-access-h5wqj") pod "64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2" (UID: "64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2"). InnerVolumeSpecName "kube-api-access-h5wqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:50:05 crc kubenswrapper[4767]: I0317 16:50:05.442386 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5wqj\" (UniqueName: \"kubernetes.io/projected/64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2-kube-api-access-h5wqj\") on node \"crc\" DevicePath \"\"" Mar 17 16:50:06 crc kubenswrapper[4767]: I0317 16:50:06.082514 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562770-njfq7" event={"ID":"64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2","Type":"ContainerDied","Data":"1d515a8aea2cd038cb60c8bd0f4963f63d7cf6cfc6d24d1737965d6c67dc0393"} Mar 17 16:50:06 crc kubenswrapper[4767]: I0317 16:50:06.082583 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d515a8aea2cd038cb60c8bd0f4963f63d7cf6cfc6d24d1737965d6c67dc0393" Mar 17 16:50:06 crc kubenswrapper[4767]: I0317 16:50:06.082665 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562770-njfq7" Mar 17 16:50:06 crc kubenswrapper[4767]: I0317 16:50:06.109276 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562764-zsncx"] Mar 17 16:50:06 crc kubenswrapper[4767]: I0317 16:50:06.135102 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562764-zsncx"] Mar 17 16:50:07 crc kubenswrapper[4767]: I0317 16:50:07.376578 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9795ce09-d4bf-46de-b861-41289e385f36" path="/var/lib/kubelet/pods/9795ce09-d4bf-46de-b861-41289e385f36/volumes" Mar 17 16:50:23 crc kubenswrapper[4767]: I0317 16:50:23.993275 4767 scope.go:117] "RemoveContainer" containerID="b4e15f78acffc90e70fef8336caa63f6518d587e5458a6f42b96f535c9dbf612" Mar 17 16:50:34 crc kubenswrapper[4767]: I0317 16:50:34.166617 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:50:34 crc kubenswrapper[4767]: I0317 16:50:34.167212 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:50:34 crc kubenswrapper[4767]: I0317 16:50:34.167273 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 16:50:34 crc kubenswrapper[4767]: I0317 16:50:34.168043 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 16:50:34 crc kubenswrapper[4767]: I0317 16:50:34.168120 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" gracePeriod=600 Mar 17 16:50:34 crc kubenswrapper[4767]: E0317 16:50:34.296021 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:50:34 crc kubenswrapper[4767]: I0317 16:50:34.850990 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" exitCode=0 Mar 17 16:50:34 crc kubenswrapper[4767]: I0317 16:50:34.851385 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216"} Mar 17 16:50:34 crc kubenswrapper[4767]: I0317 16:50:34.851434 4767 scope.go:117] "RemoveContainer" containerID="a329b73e6c74a32ee0eed732b0cfe2e5edcb9ecafe799acd643961d80496c7af" Mar 17 16:50:34 crc kubenswrapper[4767]: I0317 16:50:34.852864 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:50:34 crc kubenswrapper[4767]: E0317 16:50:34.853676 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:50:48 crc kubenswrapper[4767]: I0317 16:50:48.354975 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:50:48 crc kubenswrapper[4767]: E0317 16:50:48.355903 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:51:02 crc kubenswrapper[4767]: I0317 16:51:02.356466 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:51:02 crc kubenswrapper[4767]: E0317 16:51:02.357382 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:51:16 crc kubenswrapper[4767]: I0317 16:51:16.354570 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:51:16 crc kubenswrapper[4767]: E0317 16:51:16.355681 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:51:27 crc kubenswrapper[4767]: I0317 16:51:27.355090 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:51:27 crc kubenswrapper[4767]: E0317 16:51:27.355960 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:51:38 crc kubenswrapper[4767]: I0317 16:51:38.355024 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:51:38 crc kubenswrapper[4767]: E0317 16:51:38.355865 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:51:49 crc kubenswrapper[4767]: I0317 16:51:49.364091 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:51:49 crc kubenswrapper[4767]: E0317 16:51:49.365137 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.154599 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562772-65nvt"] Mar 17 16:52:00 crc kubenswrapper[4767]: E0317 16:52:00.155888 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2" containerName="oc" Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.155906 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2" containerName="oc" Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.156241 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2" containerName="oc" Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.157316 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562772-65nvt" Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.159990 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.160287 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.160443 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.170394 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562772-65nvt"] Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.255975 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s999d\" (UniqueName: \"kubernetes.io/projected/dba409ed-88cf-45a7-85b1-8fcc82c8abd8-kube-api-access-s999d\") pod \"auto-csr-approver-29562772-65nvt\" (UID: \"dba409ed-88cf-45a7-85b1-8fcc82c8abd8\") " pod="openshift-infra/auto-csr-approver-29562772-65nvt" Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.358699 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s999d\" (UniqueName: \"kubernetes.io/projected/dba409ed-88cf-45a7-85b1-8fcc82c8abd8-kube-api-access-s999d\") pod \"auto-csr-approver-29562772-65nvt\" (UID: \"dba409ed-88cf-45a7-85b1-8fcc82c8abd8\") " pod="openshift-infra/auto-csr-approver-29562772-65nvt" Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.386540 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s999d\" (UniqueName: \"kubernetes.io/projected/dba409ed-88cf-45a7-85b1-8fcc82c8abd8-kube-api-access-s999d\") pod \"auto-csr-approver-29562772-65nvt\" (UID: \"dba409ed-88cf-45a7-85b1-8fcc82c8abd8\") " pod="openshift-infra/auto-csr-approver-29562772-65nvt" Mar 17 16:52:00 crc kubenswrapper[4767]: I0317 16:52:00.486003 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562772-65nvt" Mar 17 16:52:01 crc kubenswrapper[4767]: I0317 16:52:01.036718 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562772-65nvt"] Mar 17 16:52:01 crc kubenswrapper[4767]: I0317 16:52:01.048476 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 16:52:01 crc kubenswrapper[4767]: I0317 16:52:01.671024 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562772-65nvt" event={"ID":"dba409ed-88cf-45a7-85b1-8fcc82c8abd8","Type":"ContainerStarted","Data":"de35764d550b516459ff13d545b4b1c2dfc1454a96162d4e39ebc2e81a7171e9"} Mar 17 16:52:03 crc kubenswrapper[4767]: I0317 16:52:03.698864 4767 generic.go:334] "Generic (PLEG): container finished" podID="dba409ed-88cf-45a7-85b1-8fcc82c8abd8" containerID="96895fe5560233d76c92200f58b32efd4a2b869d942298f9a8b441c64bb12e84" exitCode=0 Mar 17 16:52:03 crc kubenswrapper[4767]: I0317 16:52:03.698929 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562772-65nvt" event={"ID":"dba409ed-88cf-45a7-85b1-8fcc82c8abd8","Type":"ContainerDied","Data":"96895fe5560233d76c92200f58b32efd4a2b869d942298f9a8b441c64bb12e84"} Mar 17 16:52:04 crc kubenswrapper[4767]: I0317 16:52:04.354726 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:52:04 crc kubenswrapper[4767]: E0317 16:52:04.355691 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:52:05 crc kubenswrapper[4767]: I0317 16:52:05.280356 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562772-65nvt" Mar 17 16:52:05 crc kubenswrapper[4767]: I0317 16:52:05.448515 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s999d\" (UniqueName: \"kubernetes.io/projected/dba409ed-88cf-45a7-85b1-8fcc82c8abd8-kube-api-access-s999d\") pod \"dba409ed-88cf-45a7-85b1-8fcc82c8abd8\" (UID: \"dba409ed-88cf-45a7-85b1-8fcc82c8abd8\") " Mar 17 16:52:05 crc kubenswrapper[4767]: I0317 16:52:05.461236 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dba409ed-88cf-45a7-85b1-8fcc82c8abd8-kube-api-access-s999d" (OuterVolumeSpecName: "kube-api-access-s999d") pod "dba409ed-88cf-45a7-85b1-8fcc82c8abd8" (UID: "dba409ed-88cf-45a7-85b1-8fcc82c8abd8"). InnerVolumeSpecName "kube-api-access-s999d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:52:05 crc kubenswrapper[4767]: I0317 16:52:05.556859 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s999d\" (UniqueName: \"kubernetes.io/projected/dba409ed-88cf-45a7-85b1-8fcc82c8abd8-kube-api-access-s999d\") on node \"crc\" DevicePath \"\"" Mar 17 16:52:05 crc kubenswrapper[4767]: I0317 16:52:05.729875 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562772-65nvt" event={"ID":"dba409ed-88cf-45a7-85b1-8fcc82c8abd8","Type":"ContainerDied","Data":"de35764d550b516459ff13d545b4b1c2dfc1454a96162d4e39ebc2e81a7171e9"} Mar 17 16:52:05 crc kubenswrapper[4767]: I0317 16:52:05.729953 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de35764d550b516459ff13d545b4b1c2dfc1454a96162d4e39ebc2e81a7171e9" Mar 17 16:52:05 crc kubenswrapper[4767]: I0317 16:52:05.730159 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562772-65nvt" Mar 17 16:52:06 crc kubenswrapper[4767]: I0317 16:52:06.442261 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562766-wrpqm"] Mar 17 16:52:06 crc kubenswrapper[4767]: I0317 16:52:06.466656 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562766-wrpqm"] Mar 17 16:52:07 crc kubenswrapper[4767]: I0317 16:52:07.372695 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="933c4014-a9df-4fca-a5fb-ce61f79403c8" path="/var/lib/kubelet/pods/933c4014-a9df-4fca-a5fb-ce61f79403c8/volumes" Mar 17 16:52:19 crc kubenswrapper[4767]: I0317 16:52:19.366162 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:52:19 crc kubenswrapper[4767]: E0317 16:52:19.368097 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:52:24 crc kubenswrapper[4767]: I0317 16:52:24.123835 4767 scope.go:117] "RemoveContainer" containerID="d033154a465879b6aa92090e93c2d181b85605cfe09cc5cecfeb8dd0043da031" Mar 17 16:52:32 crc kubenswrapper[4767]: I0317 16:52:32.355228 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:52:32 crc kubenswrapper[4767]: E0317 16:52:32.356248 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:52:45 crc kubenswrapper[4767]: I0317 16:52:45.355542 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:52:45 crc kubenswrapper[4767]: E0317 16:52:45.356576 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:52:45 crc kubenswrapper[4767]: I0317 16:52:45.949538 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sv566"] Mar 17 16:52:45 crc kubenswrapper[4767]: E0317 16:52:45.950680 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dba409ed-88cf-45a7-85b1-8fcc82c8abd8" containerName="oc" Mar 17 16:52:45 crc kubenswrapper[4767]: I0317 16:52:45.950704 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="dba409ed-88cf-45a7-85b1-8fcc82c8abd8" containerName="oc" Mar 17 16:52:45 crc kubenswrapper[4767]: I0317 16:52:45.951075 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="dba409ed-88cf-45a7-85b1-8fcc82c8abd8" containerName="oc" Mar 17 16:52:45 crc kubenswrapper[4767]: I0317 16:52:45.953580 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:45 crc kubenswrapper[4767]: I0317 16:52:45.961967 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sv566"] Mar 17 16:52:46 crc kubenswrapper[4767]: I0317 16:52:46.060340 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-catalog-content\") pod \"redhat-operators-sv566\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:46 crc kubenswrapper[4767]: I0317 16:52:46.060794 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqq2l\" (UniqueName: \"kubernetes.io/projected/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-kube-api-access-pqq2l\") pod \"redhat-operators-sv566\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:46 crc kubenswrapper[4767]: I0317 16:52:46.060958 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-utilities\") pod \"redhat-operators-sv566\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:46 crc kubenswrapper[4767]: I0317 16:52:46.164351 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqq2l\" (UniqueName: \"kubernetes.io/projected/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-kube-api-access-pqq2l\") pod \"redhat-operators-sv566\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:46 crc kubenswrapper[4767]: I0317 16:52:46.164495 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-utilities\") pod \"redhat-operators-sv566\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:46 crc kubenswrapper[4767]: I0317 16:52:46.164667 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-catalog-content\") pod \"redhat-operators-sv566\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:46 crc kubenswrapper[4767]: I0317 16:52:46.165559 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-utilities\") pod \"redhat-operators-sv566\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:46 crc kubenswrapper[4767]: I0317 16:52:46.165728 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-catalog-content\") pod \"redhat-operators-sv566\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:46 crc kubenswrapper[4767]: I0317 16:52:46.203243 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqq2l\" (UniqueName: \"kubernetes.io/projected/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-kube-api-access-pqq2l\") pod \"redhat-operators-sv566\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:46 crc kubenswrapper[4767]: I0317 16:52:46.290805 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:46 crc kubenswrapper[4767]: I0317 16:52:46.838544 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sv566"] Mar 17 16:52:47 crc kubenswrapper[4767]: I0317 16:52:47.427495 4767 generic.go:334] "Generic (PLEG): container finished" podID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerID="5daa9a4e3e08474ca013eee78893e0f481d297d43fefde312bc1aec90b1dea32" exitCode=0 Mar 17 16:52:47 crc kubenswrapper[4767]: I0317 16:52:47.427834 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv566" event={"ID":"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa","Type":"ContainerDied","Data":"5daa9a4e3e08474ca013eee78893e0f481d297d43fefde312bc1aec90b1dea32"} Mar 17 16:52:47 crc kubenswrapper[4767]: I0317 16:52:47.428211 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv566" event={"ID":"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa","Type":"ContainerStarted","Data":"a3b28b677c2fe0495e1b6e890af31dc2348aa1a68c685f3d5affc8a807e3d657"} Mar 17 16:52:48 crc kubenswrapper[4767]: I0317 16:52:48.443769 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv566" event={"ID":"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa","Type":"ContainerStarted","Data":"f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2"} Mar 17 16:52:53 crc kubenswrapper[4767]: I0317 16:52:53.520701 4767 generic.go:334] "Generic (PLEG): container finished" podID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerID="f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2" exitCode=0 Mar 17 16:52:53 crc kubenswrapper[4767]: I0317 16:52:53.520743 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv566" event={"ID":"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa","Type":"ContainerDied","Data":"f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2"} Mar 17 16:52:54 crc kubenswrapper[4767]: I0317 16:52:54.539575 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv566" event={"ID":"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa","Type":"ContainerStarted","Data":"8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6"} Mar 17 16:52:54 crc kubenswrapper[4767]: I0317 16:52:54.578463 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sv566" podStartSLOduration=3.014524672 podStartE2EDuration="9.578397122s" podCreationTimestamp="2026-03-17 16:52:45 +0000 UTC" firstStartedPulling="2026-03-17 16:52:47.430481703 +0000 UTC m=+4558.843797750" lastFinishedPulling="2026-03-17 16:52:53.994354153 +0000 UTC m=+4565.407670200" observedRunningTime="2026-03-17 16:52:54.572560733 +0000 UTC m=+4565.985876810" watchObservedRunningTime="2026-03-17 16:52:54.578397122 +0000 UTC m=+4565.991713179" Mar 17 16:52:56 crc kubenswrapper[4767]: I0317 16:52:56.291234 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:56 crc kubenswrapper[4767]: I0317 16:52:56.291888 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:52:57 crc kubenswrapper[4767]: I0317 16:52:57.341479 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sv566" podUID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerName="registry-server" probeResult="failure" output=< Mar 17 16:52:57 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:52:57 crc kubenswrapper[4767]: > Mar 17 16:52:59 crc kubenswrapper[4767]: I0317 16:52:59.520308 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:52:59 crc kubenswrapper[4767]: E0317 16:52:59.525823 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:53:07 crc kubenswrapper[4767]: I0317 16:53:07.611509 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sv566" podUID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerName="registry-server" probeResult="failure" output=< Mar 17 16:53:07 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:53:07 crc kubenswrapper[4767]: > Mar 17 16:53:11 crc kubenswrapper[4767]: I0317 16:53:11.354656 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:53:11 crc kubenswrapper[4767]: E0317 16:53:11.355542 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:53:16 crc kubenswrapper[4767]: I0317 16:53:16.341149 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:53:16 crc kubenswrapper[4767]: I0317 16:53:16.406684 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:53:17 crc kubenswrapper[4767]: I0317 16:53:17.148528 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sv566"] Mar 17 16:53:18 crc kubenswrapper[4767]: I0317 16:53:18.383207 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sv566" podUID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerName="registry-server" containerID="cri-o://8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6" gracePeriod=2 Mar 17 16:53:18 crc kubenswrapper[4767]: I0317 16:53:18.996508 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.196024 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-utilities\") pod \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.196187 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-catalog-content\") pod \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.196309 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqq2l\" (UniqueName: \"kubernetes.io/projected/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-kube-api-access-pqq2l\") pod \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\" (UID: \"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa\") " Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.197554 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-utilities" (OuterVolumeSpecName: "utilities") pod "cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" (UID: "cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.211294 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-kube-api-access-pqq2l" (OuterVolumeSpecName: "kube-api-access-pqq2l") pod "cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" (UID: "cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa"). InnerVolumeSpecName "kube-api-access-pqq2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.300190 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.300238 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqq2l\" (UniqueName: \"kubernetes.io/projected/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-kube-api-access-pqq2l\") on node \"crc\" DevicePath \"\"" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.392194 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" (UID: "cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.397361 4767 generic.go:334] "Generic (PLEG): container finished" podID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerID="8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6" exitCode=0 Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.397412 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv566" event={"ID":"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa","Type":"ContainerDied","Data":"8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6"} Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.397460 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv566" event={"ID":"cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa","Type":"ContainerDied","Data":"a3b28b677c2fe0495e1b6e890af31dc2348aa1a68c685f3d5affc8a807e3d657"} Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.397489 4767 scope.go:117] "RemoveContainer" containerID="8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.397496 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sv566" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.406815 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.448203 4767 scope.go:117] "RemoveContainer" containerID="f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.449673 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sv566"] Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.477103 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sv566"] Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.478325 4767 scope.go:117] "RemoveContainer" containerID="5daa9a4e3e08474ca013eee78893e0f481d297d43fefde312bc1aec90b1dea32" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.546433 4767 scope.go:117] "RemoveContainer" containerID="8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6" Mar 17 16:53:19 crc kubenswrapper[4767]: E0317 16:53:19.547558 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6\": container with ID starting with 8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6 not found: ID does not exist" containerID="8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.547618 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6"} err="failed to get container status \"8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6\": rpc error: code = NotFound desc = could not find container \"8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6\": container with ID starting with 8f757a8f86cd8baff5f8d37d31c965d328c62f7320066a0f6bf7254bb26e88c6 not found: ID does not exist" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.547654 4767 scope.go:117] "RemoveContainer" containerID="f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2" Mar 17 16:53:19 crc kubenswrapper[4767]: E0317 16:53:19.548103 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2\": container with ID starting with f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2 not found: ID does not exist" containerID="f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.548136 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2"} err="failed to get container status \"f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2\": rpc error: code = NotFound desc = could not find container \"f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2\": container with ID starting with f3df8b56aa04417905555243a4782d43ac30e144b1c52378555853ba6ef900b2 not found: ID does not exist" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.548153 4767 scope.go:117] "RemoveContainer" containerID="5daa9a4e3e08474ca013eee78893e0f481d297d43fefde312bc1aec90b1dea32" Mar 17 16:53:19 crc kubenswrapper[4767]: E0317 16:53:19.548650 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5daa9a4e3e08474ca013eee78893e0f481d297d43fefde312bc1aec90b1dea32\": container with ID starting with 5daa9a4e3e08474ca013eee78893e0f481d297d43fefde312bc1aec90b1dea32 not found: ID does not exist" containerID="5daa9a4e3e08474ca013eee78893e0f481d297d43fefde312bc1aec90b1dea32" Mar 17 16:53:19 crc kubenswrapper[4767]: I0317 16:53:19.548754 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5daa9a4e3e08474ca013eee78893e0f481d297d43fefde312bc1aec90b1dea32"} err="failed to get container status \"5daa9a4e3e08474ca013eee78893e0f481d297d43fefde312bc1aec90b1dea32\": rpc error: code = NotFound desc = could not find container \"5daa9a4e3e08474ca013eee78893e0f481d297d43fefde312bc1aec90b1dea32\": container with ID starting with 5daa9a4e3e08474ca013eee78893e0f481d297d43fefde312bc1aec90b1dea32 not found: ID does not exist" Mar 17 16:53:21 crc kubenswrapper[4767]: I0317 16:53:21.371269 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" path="/var/lib/kubelet/pods/cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa/volumes" Mar 17 16:53:25 crc kubenswrapper[4767]: I0317 16:53:25.357284 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:53:25 crc kubenswrapper[4767]: E0317 16:53:25.358952 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:53:40 crc kubenswrapper[4767]: I0317 16:53:40.354279 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:53:40 crc kubenswrapper[4767]: E0317 16:53:40.355220 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:53:53 crc kubenswrapper[4767]: I0317 16:53:53.355280 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:53:53 crc kubenswrapper[4767]: E0317 16:53:53.356266 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.169386 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562774-57plp"] Mar 17 16:54:00 crc kubenswrapper[4767]: E0317 16:54:00.170564 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerName="registry-server" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.170579 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerName="registry-server" Mar 17 16:54:00 crc kubenswrapper[4767]: E0317 16:54:00.170614 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerName="extract-content" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.170620 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerName="extract-content" Mar 17 16:54:00 crc kubenswrapper[4767]: E0317 16:54:00.170649 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerName="extract-utilities" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.170659 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerName="extract-utilities" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.170923 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd81fc6c-611e-4778-8f58-dd0c2ea6a5fa" containerName="registry-server" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.171936 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562774-57plp" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.174553 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.175674 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.175781 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.189450 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562774-57plp"] Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.219189 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fft9\" (UniqueName: \"kubernetes.io/projected/f42657e8-7838-487f-82f2-f36dc5340922-kube-api-access-6fft9\") pod \"auto-csr-approver-29562774-57plp\" (UID: \"f42657e8-7838-487f-82f2-f36dc5340922\") " pod="openshift-infra/auto-csr-approver-29562774-57plp" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.322285 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fft9\" (UniqueName: \"kubernetes.io/projected/f42657e8-7838-487f-82f2-f36dc5340922-kube-api-access-6fft9\") pod \"auto-csr-approver-29562774-57plp\" (UID: \"f42657e8-7838-487f-82f2-f36dc5340922\") " pod="openshift-infra/auto-csr-approver-29562774-57plp" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.345484 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fft9\" (UniqueName: \"kubernetes.io/projected/f42657e8-7838-487f-82f2-f36dc5340922-kube-api-access-6fft9\") pod \"auto-csr-approver-29562774-57plp\" (UID: \"f42657e8-7838-487f-82f2-f36dc5340922\") " pod="openshift-infra/auto-csr-approver-29562774-57plp" Mar 17 16:54:00 crc kubenswrapper[4767]: I0317 16:54:00.501890 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562774-57plp" Mar 17 16:54:01 crc kubenswrapper[4767]: I0317 16:54:01.115986 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562774-57plp"] Mar 17 16:54:01 crc kubenswrapper[4767]: I0317 16:54:01.991441 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562774-57plp" event={"ID":"f42657e8-7838-487f-82f2-f36dc5340922","Type":"ContainerStarted","Data":"1fa1cea54b00c7fe682e58cc5c3b61390eb58d4cb563571d8d8dc0a8a2a6b199"} Mar 17 16:54:03 crc kubenswrapper[4767]: I0317 16:54:03.005853 4767 generic.go:334] "Generic (PLEG): container finished" podID="f42657e8-7838-487f-82f2-f36dc5340922" containerID="fb8a85a4359e575e2a1e1f8705de9ebf2407ff02271eb7a595a0f10b177b76d0" exitCode=0 Mar 17 16:54:03 crc kubenswrapper[4767]: I0317 16:54:03.005910 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562774-57plp" event={"ID":"f42657e8-7838-487f-82f2-f36dc5340922","Type":"ContainerDied","Data":"fb8a85a4359e575e2a1e1f8705de9ebf2407ff02271eb7a595a0f10b177b76d0"} Mar 17 16:54:04 crc kubenswrapper[4767]: I0317 16:54:04.522405 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562774-57plp" Mar 17 16:54:04 crc kubenswrapper[4767]: I0317 16:54:04.672416 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fft9\" (UniqueName: \"kubernetes.io/projected/f42657e8-7838-487f-82f2-f36dc5340922-kube-api-access-6fft9\") pod \"f42657e8-7838-487f-82f2-f36dc5340922\" (UID: \"f42657e8-7838-487f-82f2-f36dc5340922\") " Mar 17 16:54:05 crc kubenswrapper[4767]: I0317 16:54:05.030770 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562774-57plp" event={"ID":"f42657e8-7838-487f-82f2-f36dc5340922","Type":"ContainerDied","Data":"1fa1cea54b00c7fe682e58cc5c3b61390eb58d4cb563571d8d8dc0a8a2a6b199"} Mar 17 16:54:05 crc kubenswrapper[4767]: I0317 16:54:05.031104 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fa1cea54b00c7fe682e58cc5c3b61390eb58d4cb563571d8d8dc0a8a2a6b199" Mar 17 16:54:05 crc kubenswrapper[4767]: I0317 16:54:05.031239 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562774-57plp" Mar 17 16:54:05 crc kubenswrapper[4767]: I0317 16:54:05.349970 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f42657e8-7838-487f-82f2-f36dc5340922-kube-api-access-6fft9" (OuterVolumeSpecName: "kube-api-access-6fft9") pod "f42657e8-7838-487f-82f2-f36dc5340922" (UID: "f42657e8-7838-487f-82f2-f36dc5340922"). InnerVolumeSpecName "kube-api-access-6fft9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:54:05 crc kubenswrapper[4767]: I0317 16:54:05.394598 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fft9\" (UniqueName: \"kubernetes.io/projected/f42657e8-7838-487f-82f2-f36dc5340922-kube-api-access-6fft9\") on node \"crc\" DevicePath \"\"" Mar 17 16:54:05 crc kubenswrapper[4767]: I0317 16:54:05.605695 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562768-2prsp"] Mar 17 16:54:05 crc kubenswrapper[4767]: I0317 16:54:05.621457 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562768-2prsp"] Mar 17 16:54:07 crc kubenswrapper[4767]: I0317 16:54:07.355031 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:54:07 crc kubenswrapper[4767]: E0317 16:54:07.356051 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:54:07 crc kubenswrapper[4767]: I0317 16:54:07.371209 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f471d2c-947d-455c-8a41-077987d2d293" path="/var/lib/kubelet/pods/7f471d2c-947d-455c-8a41-077987d2d293/volumes" Mar 17 16:54:20 crc kubenswrapper[4767]: I0317 16:54:20.355624 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:54:20 crc kubenswrapper[4767]: E0317 16:54:20.356810 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:54:24 crc kubenswrapper[4767]: I0317 16:54:24.532170 4767 scope.go:117] "RemoveContainer" containerID="5d11702509f4ab6589b4b0bcdb90f37497ecd810d611c6652bfb9a5820ae38ff" Mar 17 16:54:35 crc kubenswrapper[4767]: I0317 16:54:35.358451 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:54:35 crc kubenswrapper[4767]: E0317 16:54:35.359408 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:54:46 crc kubenswrapper[4767]: I0317 16:54:46.354367 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:54:46 crc kubenswrapper[4767]: E0317 16:54:46.355352 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.411186 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2flzs"] Mar 17 16:54:48 crc kubenswrapper[4767]: E0317 16:54:48.415162 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f42657e8-7838-487f-82f2-f36dc5340922" containerName="oc" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.415234 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f42657e8-7838-487f-82f2-f36dc5340922" containerName="oc" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.417444 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f42657e8-7838-487f-82f2-f36dc5340922" containerName="oc" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.420269 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.437753 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2flzs"] Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.581334 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-catalog-content\") pod \"certified-operators-2flzs\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.581396 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqv7c\" (UniqueName: \"kubernetes.io/projected/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-kube-api-access-jqv7c\") pod \"certified-operators-2flzs\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.581717 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-utilities\") pod \"certified-operators-2flzs\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.684363 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-catalog-content\") pod \"certified-operators-2flzs\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.684421 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqv7c\" (UniqueName: \"kubernetes.io/projected/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-kube-api-access-jqv7c\") pod \"certified-operators-2flzs\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.684495 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-utilities\") pod \"certified-operators-2flzs\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.685071 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-catalog-content\") pod \"certified-operators-2flzs\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.685215 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-utilities\") pod \"certified-operators-2flzs\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.711998 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqv7c\" (UniqueName: \"kubernetes.io/projected/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-kube-api-access-jqv7c\") pod \"certified-operators-2flzs\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:48 crc kubenswrapper[4767]: I0317 16:54:48.746225 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:49 crc kubenswrapper[4767]: I0317 16:54:49.384983 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2flzs"] Mar 17 16:54:49 crc kubenswrapper[4767]: I0317 16:54:49.794337 4767 generic.go:334] "Generic (PLEG): container finished" podID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" containerID="c2e933d820fdf68b1c0bfac35b10215a358aed1f280fc5407683299856258ac1" exitCode=0 Mar 17 16:54:49 crc kubenswrapper[4767]: I0317 16:54:49.794469 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2flzs" event={"ID":"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9","Type":"ContainerDied","Data":"c2e933d820fdf68b1c0bfac35b10215a358aed1f280fc5407683299856258ac1"} Mar 17 16:54:49 crc kubenswrapper[4767]: I0317 16:54:49.794732 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2flzs" event={"ID":"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9","Type":"ContainerStarted","Data":"73cc658ee42324f3701063912f21c6cfbf9115aec80b82de6eb80b580a12f6b4"} Mar 17 16:54:50 crc kubenswrapper[4767]: I0317 16:54:50.823721 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2flzs" event={"ID":"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9","Type":"ContainerStarted","Data":"78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485"} Mar 17 16:54:52 crc kubenswrapper[4767]: I0317 16:54:52.859918 4767 generic.go:334] "Generic (PLEG): container finished" podID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" containerID="78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485" exitCode=0 Mar 17 16:54:52 crc kubenswrapper[4767]: I0317 16:54:52.860003 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2flzs" event={"ID":"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9","Type":"ContainerDied","Data":"78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485"} Mar 17 16:54:53 crc kubenswrapper[4767]: I0317 16:54:53.877079 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2flzs" event={"ID":"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9","Type":"ContainerStarted","Data":"b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e"} Mar 17 16:54:53 crc kubenswrapper[4767]: I0317 16:54:53.899977 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2flzs" podStartSLOduration=2.177828046 podStartE2EDuration="5.899951533s" podCreationTimestamp="2026-03-17 16:54:48 +0000 UTC" firstStartedPulling="2026-03-17 16:54:49.800015816 +0000 UTC m=+4681.213331863" lastFinishedPulling="2026-03-17 16:54:53.522139313 +0000 UTC m=+4684.935455350" observedRunningTime="2026-03-17 16:54:53.898213313 +0000 UTC m=+4685.311529380" watchObservedRunningTime="2026-03-17 16:54:53.899951533 +0000 UTC m=+4685.313267580" Mar 17 16:54:57 crc kubenswrapper[4767]: I0317 16:54:57.600724 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:54:57 crc kubenswrapper[4767]: E0317 16:54:57.602813 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:54:58 crc kubenswrapper[4767]: I0317 16:54:58.746669 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:58 crc kubenswrapper[4767]: I0317 16:54:58.747012 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:58 crc kubenswrapper[4767]: I0317 16:54:58.818500 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:58 crc kubenswrapper[4767]: I0317 16:54:58.991247 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:54:59 crc kubenswrapper[4767]: I0317 16:54:59.063833 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2flzs"] Mar 17 16:55:00 crc kubenswrapper[4767]: I0317 16:55:00.950115 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2flzs" podUID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" containerName="registry-server" containerID="cri-o://b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e" gracePeriod=2 Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.534056 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.631974 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-utilities\") pod \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.632711 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqv7c\" (UniqueName: \"kubernetes.io/projected/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-kube-api-access-jqv7c\") pod \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.632995 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-catalog-content\") pod \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\" (UID: \"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9\") " Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.633198 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-utilities" (OuterVolumeSpecName: "utilities") pod "225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" (UID: "225a2b47-3a7c-4e4a-890e-eb65cc3fcee9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.642991 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.650540 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-kube-api-access-jqv7c" (OuterVolumeSpecName: "kube-api-access-jqv7c") pod "225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" (UID: "225a2b47-3a7c-4e4a-890e-eb65cc3fcee9"). InnerVolumeSpecName "kube-api-access-jqv7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.704909 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" (UID: "225a2b47-3a7c-4e4a-890e-eb65cc3fcee9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.745792 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqv7c\" (UniqueName: \"kubernetes.io/projected/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-kube-api-access-jqv7c\") on node \"crc\" DevicePath \"\"" Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.746086 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.967080 4767 generic.go:334] "Generic (PLEG): container finished" podID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" containerID="b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e" exitCode=0 Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.967129 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2flzs" event={"ID":"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9","Type":"ContainerDied","Data":"b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e"} Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.968661 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2flzs" event={"ID":"225a2b47-3a7c-4e4a-890e-eb65cc3fcee9","Type":"ContainerDied","Data":"73cc658ee42324f3701063912f21c6cfbf9115aec80b82de6eb80b580a12f6b4"} Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.967300 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2flzs" Mar 17 16:55:01 crc kubenswrapper[4767]: I0317 16:55:01.968739 4767 scope.go:117] "RemoveContainer" containerID="b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e" Mar 17 16:55:02 crc kubenswrapper[4767]: I0317 16:55:02.006610 4767 scope.go:117] "RemoveContainer" containerID="78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485" Mar 17 16:55:02 crc kubenswrapper[4767]: I0317 16:55:02.014951 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2flzs"] Mar 17 16:55:02 crc kubenswrapper[4767]: I0317 16:55:02.028002 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2flzs"] Mar 17 16:55:02 crc kubenswrapper[4767]: I0317 16:55:02.036041 4767 scope.go:117] "RemoveContainer" containerID="c2e933d820fdf68b1c0bfac35b10215a358aed1f280fc5407683299856258ac1" Mar 17 16:55:02 crc kubenswrapper[4767]: I0317 16:55:02.091279 4767 scope.go:117] "RemoveContainer" containerID="b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e" Mar 17 16:55:02 crc kubenswrapper[4767]: E0317 16:55:02.091735 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e\": container with ID starting with b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e not found: ID does not exist" containerID="b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e" Mar 17 16:55:02 crc kubenswrapper[4767]: I0317 16:55:02.091812 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e"} err="failed to get container status \"b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e\": rpc error: code = NotFound desc = could not find container \"b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e\": container with ID starting with b7675b039a76f56e6bc13334fc52b8cd75b65fc93452e3e4c65d74f455bedf7e not found: ID does not exist" Mar 17 16:55:02 crc kubenswrapper[4767]: I0317 16:55:02.091887 4767 scope.go:117] "RemoveContainer" containerID="78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485" Mar 17 16:55:02 crc kubenswrapper[4767]: E0317 16:55:02.092156 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485\": container with ID starting with 78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485 not found: ID does not exist" containerID="78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485" Mar 17 16:55:02 crc kubenswrapper[4767]: I0317 16:55:02.092277 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485"} err="failed to get container status \"78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485\": rpc error: code = NotFound desc = could not find container \"78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485\": container with ID starting with 78a2da7813053674b1ad90be344dc9ac005a65392da737c16b4b76775b10f485 not found: ID does not exist" Mar 17 16:55:02 crc kubenswrapper[4767]: I0317 16:55:02.092296 4767 scope.go:117] "RemoveContainer" containerID="c2e933d820fdf68b1c0bfac35b10215a358aed1f280fc5407683299856258ac1" Mar 17 16:55:02 crc kubenswrapper[4767]: E0317 16:55:02.092565 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2e933d820fdf68b1c0bfac35b10215a358aed1f280fc5407683299856258ac1\": container with ID starting with c2e933d820fdf68b1c0bfac35b10215a358aed1f280fc5407683299856258ac1 not found: ID does not exist" containerID="c2e933d820fdf68b1c0bfac35b10215a358aed1f280fc5407683299856258ac1" Mar 17 16:55:02 crc kubenswrapper[4767]: I0317 16:55:02.092607 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2e933d820fdf68b1c0bfac35b10215a358aed1f280fc5407683299856258ac1"} err="failed to get container status \"c2e933d820fdf68b1c0bfac35b10215a358aed1f280fc5407683299856258ac1\": rpc error: code = NotFound desc = could not find container \"c2e933d820fdf68b1c0bfac35b10215a358aed1f280fc5407683299856258ac1\": container with ID starting with c2e933d820fdf68b1c0bfac35b10215a358aed1f280fc5407683299856258ac1 not found: ID does not exist" Mar 17 16:55:03 crc kubenswrapper[4767]: I0317 16:55:03.371839 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" path="/var/lib/kubelet/pods/225a2b47-3a7c-4e4a-890e-eb65cc3fcee9/volumes" Mar 17 16:55:10 crc kubenswrapper[4767]: I0317 16:55:10.354354 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:55:10 crc kubenswrapper[4767]: E0317 16:55:10.355300 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:55:24 crc kubenswrapper[4767]: I0317 16:55:24.355194 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:55:24 crc kubenswrapper[4767]: E0317 16:55:24.356029 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 16:55:35 crc kubenswrapper[4767]: I0317 16:55:35.356217 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 16:55:36 crc kubenswrapper[4767]: I0317 16:55:36.395691 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"c4883088f1ca02db880affb212da5ef91b29d53adeb7d812b4a4c5c2a0d946f2"} Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.151715 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562776-dt5ql"] Mar 17 16:56:00 crc kubenswrapper[4767]: E0317 16:56:00.152831 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" containerName="extract-content" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.152845 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" containerName="extract-content" Mar 17 16:56:00 crc kubenswrapper[4767]: E0317 16:56:00.152894 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" containerName="extract-utilities" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.152902 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" containerName="extract-utilities" Mar 17 16:56:00 crc kubenswrapper[4767]: E0317 16:56:00.152937 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" containerName="registry-server" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.152944 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" containerName="registry-server" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.153250 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="225a2b47-3a7c-4e4a-890e-eb65cc3fcee9" containerName="registry-server" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.154149 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562776-dt5ql" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.156390 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.156471 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.156533 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.175786 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562776-dt5ql"] Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.258770 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5tg7\" (UniqueName: \"kubernetes.io/projected/a8bcb562-f3bf-4856-bbeb-b67ed48d147f-kube-api-access-c5tg7\") pod \"auto-csr-approver-29562776-dt5ql\" (UID: \"a8bcb562-f3bf-4856-bbeb-b67ed48d147f\") " pod="openshift-infra/auto-csr-approver-29562776-dt5ql" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.361156 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5tg7\" (UniqueName: \"kubernetes.io/projected/a8bcb562-f3bf-4856-bbeb-b67ed48d147f-kube-api-access-c5tg7\") pod \"auto-csr-approver-29562776-dt5ql\" (UID: \"a8bcb562-f3bf-4856-bbeb-b67ed48d147f\") " pod="openshift-infra/auto-csr-approver-29562776-dt5ql" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.384965 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5tg7\" (UniqueName: \"kubernetes.io/projected/a8bcb562-f3bf-4856-bbeb-b67ed48d147f-kube-api-access-c5tg7\") pod \"auto-csr-approver-29562776-dt5ql\" (UID: \"a8bcb562-f3bf-4856-bbeb-b67ed48d147f\") " pod="openshift-infra/auto-csr-approver-29562776-dt5ql" Mar 17 16:56:00 crc kubenswrapper[4767]: I0317 16:56:00.495157 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562776-dt5ql" Mar 17 16:56:01 crc kubenswrapper[4767]: I0317 16:56:01.000780 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562776-dt5ql"] Mar 17 16:56:01 crc kubenswrapper[4767]: I0317 16:56:01.686975 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562776-dt5ql" event={"ID":"a8bcb562-f3bf-4856-bbeb-b67ed48d147f","Type":"ContainerStarted","Data":"9e3cd1b0b9244feaeaf915138ec1f79bde1b0a8c251af0d335e9a264acd33927"} Mar 17 16:56:02 crc kubenswrapper[4767]: I0317 16:56:02.722931 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562776-dt5ql" event={"ID":"a8bcb562-f3bf-4856-bbeb-b67ed48d147f","Type":"ContainerStarted","Data":"7946dbc889642db464ce137528a6663b61012af67d1fc270514467ca79680c66"} Mar 17 16:56:02 crc kubenswrapper[4767]: I0317 16:56:02.761403 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562776-dt5ql" podStartSLOduration=1.7821026020000001 podStartE2EDuration="2.76134446s" podCreationTimestamp="2026-03-17 16:56:00 +0000 UTC" firstStartedPulling="2026-03-17 16:56:00.998830752 +0000 UTC m=+4752.412146799" lastFinishedPulling="2026-03-17 16:56:01.97807261 +0000 UTC m=+4753.391388657" observedRunningTime="2026-03-17 16:56:02.742665796 +0000 UTC m=+4754.155981863" watchObservedRunningTime="2026-03-17 16:56:02.76134446 +0000 UTC m=+4754.174660507" Mar 17 16:56:03 crc kubenswrapper[4767]: I0317 16:56:03.736504 4767 generic.go:334] "Generic (PLEG): container finished" podID="a8bcb562-f3bf-4856-bbeb-b67ed48d147f" containerID="7946dbc889642db464ce137528a6663b61012af67d1fc270514467ca79680c66" exitCode=0 Mar 17 16:56:03 crc kubenswrapper[4767]: I0317 16:56:03.737253 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562776-dt5ql" event={"ID":"a8bcb562-f3bf-4856-bbeb-b67ed48d147f","Type":"ContainerDied","Data":"7946dbc889642db464ce137528a6663b61012af67d1fc270514467ca79680c66"} Mar 17 16:56:05 crc kubenswrapper[4767]: I0317 16:56:05.495910 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562776-dt5ql" Mar 17 16:56:05 crc kubenswrapper[4767]: I0317 16:56:05.683817 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5tg7\" (UniqueName: \"kubernetes.io/projected/a8bcb562-f3bf-4856-bbeb-b67ed48d147f-kube-api-access-c5tg7\") pod \"a8bcb562-f3bf-4856-bbeb-b67ed48d147f\" (UID: \"a8bcb562-f3bf-4856-bbeb-b67ed48d147f\") " Mar 17 16:56:05 crc kubenswrapper[4767]: I0317 16:56:05.692038 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8bcb562-f3bf-4856-bbeb-b67ed48d147f-kube-api-access-c5tg7" (OuterVolumeSpecName: "kube-api-access-c5tg7") pod "a8bcb562-f3bf-4856-bbeb-b67ed48d147f" (UID: "a8bcb562-f3bf-4856-bbeb-b67ed48d147f"). InnerVolumeSpecName "kube-api-access-c5tg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:56:05 crc kubenswrapper[4767]: I0317 16:56:05.758760 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562776-dt5ql" event={"ID":"a8bcb562-f3bf-4856-bbeb-b67ed48d147f","Type":"ContainerDied","Data":"9e3cd1b0b9244feaeaf915138ec1f79bde1b0a8c251af0d335e9a264acd33927"} Mar 17 16:56:05 crc kubenswrapper[4767]: I0317 16:56:05.758810 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e3cd1b0b9244feaeaf915138ec1f79bde1b0a8c251af0d335e9a264acd33927" Mar 17 16:56:05 crc kubenswrapper[4767]: I0317 16:56:05.758855 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562776-dt5ql" Mar 17 16:56:05 crc kubenswrapper[4767]: I0317 16:56:05.791290 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5tg7\" (UniqueName: \"kubernetes.io/projected/a8bcb562-f3bf-4856-bbeb-b67ed48d147f-kube-api-access-c5tg7\") on node \"crc\" DevicePath \"\"" Mar 17 16:56:05 crc kubenswrapper[4767]: I0317 16:56:05.827866 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562770-njfq7"] Mar 17 16:56:05 crc kubenswrapper[4767]: I0317 16:56:05.840913 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562770-njfq7"] Mar 17 16:56:07 crc kubenswrapper[4767]: I0317 16:56:07.367451 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2" path="/var/lib/kubelet/pods/64cb8ff1-d1e8-4aab-be4a-3b5716a8faa2/volumes" Mar 17 16:56:24 crc kubenswrapper[4767]: I0317 16:56:24.691194 4767 scope.go:117] "RemoveContainer" containerID="64888c8523b4e039327ffdddd6972f0fdb2ecd8400989b45bd137355fc3ce45d" Mar 17 16:56:34 crc kubenswrapper[4767]: I0317 16:56:34.888059 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l6tgf"] Mar 17 16:56:34 crc kubenswrapper[4767]: E0317 16:56:34.889212 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8bcb562-f3bf-4856-bbeb-b67ed48d147f" containerName="oc" Mar 17 16:56:34 crc kubenswrapper[4767]: I0317 16:56:34.889228 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8bcb562-f3bf-4856-bbeb-b67ed48d147f" containerName="oc" Mar 17 16:56:34 crc kubenswrapper[4767]: I0317 16:56:34.889488 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8bcb562-f3bf-4856-bbeb-b67ed48d147f" containerName="oc" Mar 17 16:56:34 crc kubenswrapper[4767]: I0317 16:56:34.891505 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:34 crc kubenswrapper[4767]: I0317 16:56:34.911400 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l6tgf"] Mar 17 16:56:35 crc kubenswrapper[4767]: I0317 16:56:35.069850 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68khx\" (UniqueName: \"kubernetes.io/projected/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-kube-api-access-68khx\") pod \"redhat-marketplace-l6tgf\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:35 crc kubenswrapper[4767]: I0317 16:56:35.069996 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-utilities\") pod \"redhat-marketplace-l6tgf\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:35 crc kubenswrapper[4767]: I0317 16:56:35.070235 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-catalog-content\") pod \"redhat-marketplace-l6tgf\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:35 crc kubenswrapper[4767]: I0317 16:56:35.173232 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68khx\" (UniqueName: \"kubernetes.io/projected/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-kube-api-access-68khx\") pod \"redhat-marketplace-l6tgf\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:35 crc kubenswrapper[4767]: I0317 16:56:35.173608 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-utilities\") pod \"redhat-marketplace-l6tgf\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:35 crc kubenswrapper[4767]: I0317 16:56:35.173685 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-catalog-content\") pod \"redhat-marketplace-l6tgf\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:35 crc kubenswrapper[4767]: I0317 16:56:35.174147 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-catalog-content\") pod \"redhat-marketplace-l6tgf\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:35 crc kubenswrapper[4767]: I0317 16:56:35.174237 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-utilities\") pod \"redhat-marketplace-l6tgf\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:35 crc kubenswrapper[4767]: I0317 16:56:35.196215 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68khx\" (UniqueName: \"kubernetes.io/projected/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-kube-api-access-68khx\") pod \"redhat-marketplace-l6tgf\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:35 crc kubenswrapper[4767]: I0317 16:56:35.227157 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:35 crc kubenswrapper[4767]: I0317 16:56:35.849847 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l6tgf"] Mar 17 16:56:36 crc kubenswrapper[4767]: I0317 16:56:36.139953 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l6tgf" event={"ID":"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd","Type":"ContainerStarted","Data":"700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5"} Mar 17 16:56:36 crc kubenswrapper[4767]: I0317 16:56:36.140314 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l6tgf" event={"ID":"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd","Type":"ContainerStarted","Data":"b5039512e12038e3e683a6f427bfd50d61e87b9f41b24729dc2fec915a50e952"} Mar 17 16:56:37 crc kubenswrapper[4767]: I0317 16:56:37.152186 4767 generic.go:334] "Generic (PLEG): container finished" podID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" containerID="700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5" exitCode=0 Mar 17 16:56:37 crc kubenswrapper[4767]: I0317 16:56:37.152307 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l6tgf" event={"ID":"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd","Type":"ContainerDied","Data":"700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5"} Mar 17 16:56:38 crc kubenswrapper[4767]: I0317 16:56:38.169765 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l6tgf" event={"ID":"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd","Type":"ContainerStarted","Data":"369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d"} Mar 17 16:56:39 crc kubenswrapper[4767]: I0317 16:56:39.184470 4767 generic.go:334] "Generic (PLEG): container finished" podID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" containerID="369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d" exitCode=0 Mar 17 16:56:39 crc kubenswrapper[4767]: I0317 16:56:39.184780 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l6tgf" event={"ID":"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd","Type":"ContainerDied","Data":"369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d"} Mar 17 16:56:41 crc kubenswrapper[4767]: I0317 16:56:41.243908 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l6tgf" event={"ID":"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd","Type":"ContainerStarted","Data":"a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d"} Mar 17 16:56:41 crc kubenswrapper[4767]: I0317 16:56:41.279551 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l6tgf" podStartSLOduration=3.68085242 podStartE2EDuration="7.27952433s" podCreationTimestamp="2026-03-17 16:56:34 +0000 UTC" firstStartedPulling="2026-03-17 16:56:37.155012055 +0000 UTC m=+4788.568328102" lastFinishedPulling="2026-03-17 16:56:40.753683965 +0000 UTC m=+4792.167000012" observedRunningTime="2026-03-17 16:56:41.266713876 +0000 UTC m=+4792.680029953" watchObservedRunningTime="2026-03-17 16:56:41.27952433 +0000 UTC m=+4792.692840377" Mar 17 16:56:42 crc kubenswrapper[4767]: I0317 16:56:42.991048 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7rrdz"] Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.005942 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.063792 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-utilities\") pod \"community-operators-7rrdz\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.064229 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-catalog-content\") pod \"community-operators-7rrdz\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.064278 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp7v8\" (UniqueName: \"kubernetes.io/projected/8d6f284e-81c0-4754-a09a-d72df8474941-kube-api-access-jp7v8\") pod \"community-operators-7rrdz\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.068973 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7rrdz"] Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.167201 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-utilities\") pod \"community-operators-7rrdz\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.167252 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-catalog-content\") pod \"community-operators-7rrdz\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.167293 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp7v8\" (UniqueName: \"kubernetes.io/projected/8d6f284e-81c0-4754-a09a-d72df8474941-kube-api-access-jp7v8\") pod \"community-operators-7rrdz\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.168086 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-catalog-content\") pod \"community-operators-7rrdz\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.168086 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-utilities\") pod \"community-operators-7rrdz\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.322760 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp7v8\" (UniqueName: \"kubernetes.io/projected/8d6f284e-81c0-4754-a09a-d72df8474941-kube-api-access-jp7v8\") pod \"community-operators-7rrdz\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:43 crc kubenswrapper[4767]: I0317 16:56:43.356531 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:44 crc kubenswrapper[4767]: I0317 16:56:44.267233 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7rrdz"] Mar 17 16:56:44 crc kubenswrapper[4767]: I0317 16:56:44.389974 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rrdz" event={"ID":"8d6f284e-81c0-4754-a09a-d72df8474941","Type":"ContainerStarted","Data":"bc42b6a87d0752cbc698eb82a35ca3efbaa29ce79debabb8ad54a227852f6866"} Mar 17 16:56:45 crc kubenswrapper[4767]: I0317 16:56:45.228304 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:45 crc kubenswrapper[4767]: I0317 16:56:45.228376 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:45 crc kubenswrapper[4767]: I0317 16:56:45.412522 4767 generic.go:334] "Generic (PLEG): container finished" podID="8d6f284e-81c0-4754-a09a-d72df8474941" containerID="07ea63fd510e4a6392b7080d150b701788dd3afe09b95e0e59db083668f47101" exitCode=0 Mar 17 16:56:45 crc kubenswrapper[4767]: I0317 16:56:45.412584 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rrdz" event={"ID":"8d6f284e-81c0-4754-a09a-d72df8474941","Type":"ContainerDied","Data":"07ea63fd510e4a6392b7080d150b701788dd3afe09b95e0e59db083668f47101"} Mar 17 16:56:45 crc kubenswrapper[4767]: I0317 16:56:45.703761 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:45 crc kubenswrapper[4767]: I0317 16:56:45.765327 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:47 crc kubenswrapper[4767]: I0317 16:56:47.447845 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rrdz" event={"ID":"8d6f284e-81c0-4754-a09a-d72df8474941","Type":"ContainerStarted","Data":"e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373"} Mar 17 16:56:48 crc kubenswrapper[4767]: I0317 16:56:48.461374 4767 generic.go:334] "Generic (PLEG): container finished" podID="8d6f284e-81c0-4754-a09a-d72df8474941" containerID="e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373" exitCode=0 Mar 17 16:56:48 crc kubenswrapper[4767]: I0317 16:56:48.461431 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rrdz" event={"ID":"8d6f284e-81c0-4754-a09a-d72df8474941","Type":"ContainerDied","Data":"e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373"} Mar 17 16:56:48 crc kubenswrapper[4767]: I0317 16:56:48.630588 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l6tgf"] Mar 17 16:56:48 crc kubenswrapper[4767]: I0317 16:56:48.631258 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l6tgf" podUID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" containerName="registry-server" containerID="cri-o://a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d" gracePeriod=2 Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.414340 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.519633 4767 generic.go:334] "Generic (PLEG): container finished" podID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" containerID="a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d" exitCode=0 Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.520026 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l6tgf" event={"ID":"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd","Type":"ContainerDied","Data":"a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d"} Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.520063 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l6tgf" event={"ID":"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd","Type":"ContainerDied","Data":"b5039512e12038e3e683a6f427bfd50d61e87b9f41b24729dc2fec915a50e952"} Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.520087 4767 scope.go:117] "RemoveContainer" containerID="a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.520333 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l6tgf" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.559349 4767 scope.go:117] "RemoveContainer" containerID="369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.561104 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68khx\" (UniqueName: \"kubernetes.io/projected/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-kube-api-access-68khx\") pod \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.561275 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-catalog-content\") pod \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.561472 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-utilities\") pod \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\" (UID: \"3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd\") " Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.562912 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-utilities" (OuterVolumeSpecName: "utilities") pod "3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" (UID: "3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.573742 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-kube-api-access-68khx" (OuterVolumeSpecName: "kube-api-access-68khx") pod "3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" (UID: "3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd"). InnerVolumeSpecName "kube-api-access-68khx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.593334 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" (UID: "3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.603337 4767 scope.go:117] "RemoveContainer" containerID="700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.664631 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68khx\" (UniqueName: \"kubernetes.io/projected/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-kube-api-access-68khx\") on node \"crc\" DevicePath \"\"" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.664680 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.664690 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.728017 4767 scope.go:117] "RemoveContainer" containerID="a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d" Mar 17 16:56:49 crc kubenswrapper[4767]: E0317 16:56:49.728556 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d\": container with ID starting with a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d not found: ID does not exist" containerID="a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.728594 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d"} err="failed to get container status \"a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d\": rpc error: code = NotFound desc = could not find container \"a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d\": container with ID starting with a2bc51975c5038df2106a48a23003afca438a32859504a6cb12d29b8b490c61d not found: ID does not exist" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.728618 4767 scope.go:117] "RemoveContainer" containerID="369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d" Mar 17 16:56:49 crc kubenswrapper[4767]: E0317 16:56:49.729294 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d\": container with ID starting with 369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d not found: ID does not exist" containerID="369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.729323 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d"} err="failed to get container status \"369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d\": rpc error: code = NotFound desc = could not find container \"369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d\": container with ID starting with 369016495e8d60a571f88c7fa2cbde1841a2454085bf880d1a68c9e9f30f383d not found: ID does not exist" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.729347 4767 scope.go:117] "RemoveContainer" containerID="700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5" Mar 17 16:56:49 crc kubenswrapper[4767]: E0317 16:56:49.730452 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5\": container with ID starting with 700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5 not found: ID does not exist" containerID="700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.730506 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5"} err="failed to get container status \"700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5\": rpc error: code = NotFound desc = could not find container \"700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5\": container with ID starting with 700bf0522a63ae84e0ab0444b4e0da03a1ad53cc9ae6ba49a0cfa5da807b27c5 not found: ID does not exist" Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.864526 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l6tgf"] Mar 17 16:56:49 crc kubenswrapper[4767]: I0317 16:56:49.879734 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l6tgf"] Mar 17 16:56:50 crc kubenswrapper[4767]: I0317 16:56:50.536745 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rrdz" event={"ID":"8d6f284e-81c0-4754-a09a-d72df8474941","Type":"ContainerStarted","Data":"038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da"} Mar 17 16:56:50 crc kubenswrapper[4767]: I0317 16:56:50.570265 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7rrdz" podStartSLOduration=4.949702191 podStartE2EDuration="8.570232899s" podCreationTimestamp="2026-03-17 16:56:42 +0000 UTC" firstStartedPulling="2026-03-17 16:56:45.415327262 +0000 UTC m=+4796.828643309" lastFinishedPulling="2026-03-17 16:56:49.03585797 +0000 UTC m=+4800.449174017" observedRunningTime="2026-03-17 16:56:50.559655559 +0000 UTC m=+4801.972971626" watchObservedRunningTime="2026-03-17 16:56:50.570232899 +0000 UTC m=+4801.983548946" Mar 17 16:56:51 crc kubenswrapper[4767]: I0317 16:56:51.370109 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" path="/var/lib/kubelet/pods/3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd/volumes" Mar 17 16:56:53 crc kubenswrapper[4767]: I0317 16:56:53.377242 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:53 crc kubenswrapper[4767]: I0317 16:56:53.377927 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:56:54 crc kubenswrapper[4767]: I0317 16:56:54.417587 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-7rrdz" podUID="8d6f284e-81c0-4754-a09a-d72df8474941" containerName="registry-server" probeResult="failure" output=< Mar 17 16:56:54 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 16:56:54 crc kubenswrapper[4767]: > Mar 17 16:57:03 crc kubenswrapper[4767]: I0317 16:57:03.414834 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:57:03 crc kubenswrapper[4767]: I0317 16:57:03.489080 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:57:03 crc kubenswrapper[4767]: I0317 16:57:03.660440 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7rrdz"] Mar 17 16:57:04 crc kubenswrapper[4767]: I0317 16:57:04.915063 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7rrdz" podUID="8d6f284e-81c0-4754-a09a-d72df8474941" containerName="registry-server" containerID="cri-o://038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da" gracePeriod=2 Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.696764 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.832444 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-catalog-content\") pod \"8d6f284e-81c0-4754-a09a-d72df8474941\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.832835 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-utilities\") pod \"8d6f284e-81c0-4754-a09a-d72df8474941\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.833123 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp7v8\" (UniqueName: \"kubernetes.io/projected/8d6f284e-81c0-4754-a09a-d72df8474941-kube-api-access-jp7v8\") pod \"8d6f284e-81c0-4754-a09a-d72df8474941\" (UID: \"8d6f284e-81c0-4754-a09a-d72df8474941\") " Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.834757 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-utilities" (OuterVolumeSpecName: "utilities") pod "8d6f284e-81c0-4754-a09a-d72df8474941" (UID: "8d6f284e-81c0-4754-a09a-d72df8474941"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.841298 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d6f284e-81c0-4754-a09a-d72df8474941-kube-api-access-jp7v8" (OuterVolumeSpecName: "kube-api-access-jp7v8") pod "8d6f284e-81c0-4754-a09a-d72df8474941" (UID: "8d6f284e-81c0-4754-a09a-d72df8474941"). InnerVolumeSpecName "kube-api-access-jp7v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.857802 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.857867 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp7v8\" (UniqueName: \"kubernetes.io/projected/8d6f284e-81c0-4754-a09a-d72df8474941-kube-api-access-jp7v8\") on node \"crc\" DevicePath \"\"" Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.925916 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d6f284e-81c0-4754-a09a-d72df8474941" (UID: "8d6f284e-81c0-4754-a09a-d72df8474941"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.935259 4767 generic.go:334] "Generic (PLEG): container finished" podID="8d6f284e-81c0-4754-a09a-d72df8474941" containerID="038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da" exitCode=0 Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.935350 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rrdz" event={"ID":"8d6f284e-81c0-4754-a09a-d72df8474941","Type":"ContainerDied","Data":"038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da"} Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.935413 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rrdz" event={"ID":"8d6f284e-81c0-4754-a09a-d72df8474941","Type":"ContainerDied","Data":"bc42b6a87d0752cbc698eb82a35ca3efbaa29ce79debabb8ad54a227852f6866"} Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.935434 4767 scope.go:117] "RemoveContainer" containerID="038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da" Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.935750 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7rrdz" Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.961815 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6f284e-81c0-4754-a09a-d72df8474941-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.970939 4767 scope.go:117] "RemoveContainer" containerID="e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373" Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.985030 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7rrdz"] Mar 17 16:57:05 crc kubenswrapper[4767]: I0317 16:57:05.999523 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7rrdz"] Mar 17 16:57:06 crc kubenswrapper[4767]: I0317 16:57:06.012300 4767 scope.go:117] "RemoveContainer" containerID="07ea63fd510e4a6392b7080d150b701788dd3afe09b95e0e59db083668f47101" Mar 17 16:57:06 crc kubenswrapper[4767]: I0317 16:57:06.116535 4767 scope.go:117] "RemoveContainer" containerID="038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da" Mar 17 16:57:06 crc kubenswrapper[4767]: E0317 16:57:06.117397 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da\": container with ID starting with 038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da not found: ID does not exist" containerID="038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da" Mar 17 16:57:06 crc kubenswrapper[4767]: I0317 16:57:06.117445 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da"} err="failed to get container status \"038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da\": rpc error: code = NotFound desc = could not find container \"038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da\": container with ID starting with 038e9370eafcedafc2186644432f3dcfdbc230d3ad3ee5a13cdfb4552affe7da not found: ID does not exist" Mar 17 16:57:06 crc kubenswrapper[4767]: I0317 16:57:06.117479 4767 scope.go:117] "RemoveContainer" containerID="e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373" Mar 17 16:57:06 crc kubenswrapper[4767]: E0317 16:57:06.118087 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373\": container with ID starting with e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373 not found: ID does not exist" containerID="e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373" Mar 17 16:57:06 crc kubenswrapper[4767]: I0317 16:57:06.118512 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373"} err="failed to get container status \"e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373\": rpc error: code = NotFound desc = could not find container \"e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373\": container with ID starting with e89517c20bb0997c3733dc19207ba3f8eff1c03d7f4d8d7b3564f06625ade373 not found: ID does not exist" Mar 17 16:57:06 crc kubenswrapper[4767]: I0317 16:57:06.118632 4767 scope.go:117] "RemoveContainer" containerID="07ea63fd510e4a6392b7080d150b701788dd3afe09b95e0e59db083668f47101" Mar 17 16:57:06 crc kubenswrapper[4767]: E0317 16:57:06.124416 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ea63fd510e4a6392b7080d150b701788dd3afe09b95e0e59db083668f47101\": container with ID starting with 07ea63fd510e4a6392b7080d150b701788dd3afe09b95e0e59db083668f47101 not found: ID does not exist" containerID="07ea63fd510e4a6392b7080d150b701788dd3afe09b95e0e59db083668f47101" Mar 17 16:57:06 crc kubenswrapper[4767]: I0317 16:57:06.124477 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ea63fd510e4a6392b7080d150b701788dd3afe09b95e0e59db083668f47101"} err="failed to get container status \"07ea63fd510e4a6392b7080d150b701788dd3afe09b95e0e59db083668f47101\": rpc error: code = NotFound desc = could not find container \"07ea63fd510e4a6392b7080d150b701788dd3afe09b95e0e59db083668f47101\": container with ID starting with 07ea63fd510e4a6392b7080d150b701788dd3afe09b95e0e59db083668f47101 not found: ID does not exist" Mar 17 16:57:07 crc kubenswrapper[4767]: I0317 16:57:07.368605 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d6f284e-81c0-4754-a09a-d72df8474941" path="/var/lib/kubelet/pods/8d6f284e-81c0-4754-a09a-d72df8474941/volumes" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.375301 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Mar 17 16:57:55 crc kubenswrapper[4767]: E0317 16:57:55.378907 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6f284e-81c0-4754-a09a-d72df8474941" containerName="registry-server" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.378929 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6f284e-81c0-4754-a09a-d72df8474941" containerName="registry-server" Mar 17 16:57:55 crc kubenswrapper[4767]: E0317 16:57:55.378938 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6f284e-81c0-4754-a09a-d72df8474941" containerName="extract-utilities" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.378947 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6f284e-81c0-4754-a09a-d72df8474941" containerName="extract-utilities" Mar 17 16:57:55 crc kubenswrapper[4767]: E0317 16:57:55.378966 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" containerName="registry-server" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.378972 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" containerName="registry-server" Mar 17 16:57:55 crc kubenswrapper[4767]: E0317 16:57:55.378985 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" containerName="extract-content" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.378990 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" containerName="extract-content" Mar 17 16:57:55 crc kubenswrapper[4767]: E0317 16:57:55.379004 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6f284e-81c0-4754-a09a-d72df8474941" containerName="extract-content" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.379010 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6f284e-81c0-4754-a09a-d72df8474941" containerName="extract-content" Mar 17 16:57:55 crc kubenswrapper[4767]: E0317 16:57:55.379037 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" containerName="extract-utilities" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.379045 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" containerName="extract-utilities" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.379412 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dbed36b-1ea7-4cad-96ed-c1b724a0e8dd" containerName="registry-server" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.379431 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d6f284e-81c0-4754-a09a-d72df8474941" containerName="registry-server" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.381332 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.381529 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.392066 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.392298 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.392605 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-c8ndj" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.392803 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.421692 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.422143 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.422262 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.422295 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-config-data\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.422400 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4zw7\" (UniqueName: \"kubernetes.io/projected/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-kube-api-access-k4zw7\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.422463 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.422536 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.422576 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.422631 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.524429 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.524555 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-config-data\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.524648 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4zw7\" (UniqueName: \"kubernetes.io/projected/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-kube-api-access-k4zw7\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.524705 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.524770 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.524803 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.524832 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.524870 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.524924 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.532415 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.532932 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-config-data\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.533296 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.533834 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.534029 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.534293 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.535557 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.537037 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.555310 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4zw7\" (UniqueName: \"kubernetes.io/projected/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-kube-api-access-k4zw7\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.596436 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " pod="openstack/tempest-tests-tempest" Mar 17 16:57:55 crc kubenswrapper[4767]: I0317 16:57:55.774819 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 17 16:57:56 crc kubenswrapper[4767]: I0317 16:57:56.353870 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 17 16:57:56 crc kubenswrapper[4767]: I0317 16:57:56.367651 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 16:57:57 crc kubenswrapper[4767]: I0317 16:57:57.265599 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"909f6b35-823c-4b1b-ba3b-6db1a4f6b444","Type":"ContainerStarted","Data":"e3039e512befc9505271561e6ed0eeab8562ce4f3543358b67aa0d2b146eb99c"} Mar 17 16:58:00 crc kubenswrapper[4767]: I0317 16:58:00.163685 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562778-5cxf5"] Mar 17 16:58:00 crc kubenswrapper[4767]: I0317 16:58:00.167133 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562778-5cxf5" Mar 17 16:58:00 crc kubenswrapper[4767]: I0317 16:58:00.173239 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 16:58:00 crc kubenswrapper[4767]: I0317 16:58:00.173529 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 16:58:00 crc kubenswrapper[4767]: I0317 16:58:00.173789 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 16:58:00 crc kubenswrapper[4767]: I0317 16:58:00.182773 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562778-5cxf5"] Mar 17 16:58:00 crc kubenswrapper[4767]: I0317 16:58:00.313982 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94jpz\" (UniqueName: \"kubernetes.io/projected/f71df388-7598-46d0-82f6-5a6aa85d90ad-kube-api-access-94jpz\") pod \"auto-csr-approver-29562778-5cxf5\" (UID: \"f71df388-7598-46d0-82f6-5a6aa85d90ad\") " pod="openshift-infra/auto-csr-approver-29562778-5cxf5" Mar 17 16:58:00 crc kubenswrapper[4767]: I0317 16:58:00.418295 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94jpz\" (UniqueName: \"kubernetes.io/projected/f71df388-7598-46d0-82f6-5a6aa85d90ad-kube-api-access-94jpz\") pod \"auto-csr-approver-29562778-5cxf5\" (UID: \"f71df388-7598-46d0-82f6-5a6aa85d90ad\") " pod="openshift-infra/auto-csr-approver-29562778-5cxf5" Mar 17 16:58:00 crc kubenswrapper[4767]: I0317 16:58:00.439163 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94jpz\" (UniqueName: \"kubernetes.io/projected/f71df388-7598-46d0-82f6-5a6aa85d90ad-kube-api-access-94jpz\") pod \"auto-csr-approver-29562778-5cxf5\" (UID: \"f71df388-7598-46d0-82f6-5a6aa85d90ad\") " pod="openshift-infra/auto-csr-approver-29562778-5cxf5" Mar 17 16:58:00 crc kubenswrapper[4767]: I0317 16:58:00.516711 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562778-5cxf5" Mar 17 16:58:04 crc kubenswrapper[4767]: I0317 16:58:04.167037 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:58:04 crc kubenswrapper[4767]: I0317 16:58:04.167722 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:58:07 crc kubenswrapper[4767]: I0317 16:58:07.296377 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562778-5cxf5"] Mar 17 16:58:07 crc kubenswrapper[4767]: I0317 16:58:07.766468 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562778-5cxf5" event={"ID":"f71df388-7598-46d0-82f6-5a6aa85d90ad","Type":"ContainerStarted","Data":"8e0e7008802a8282bf68e1ee980a10c49c11ec86a475e334d6c4cac341e5c918"} Mar 17 16:58:09 crc kubenswrapper[4767]: I0317 16:58:09.798379 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562778-5cxf5" event={"ID":"f71df388-7598-46d0-82f6-5a6aa85d90ad","Type":"ContainerStarted","Data":"f752ad2aed374732f44f8dca7dc468c6deeb94183f06dfb477226dabf512e86a"} Mar 17 16:58:09 crc kubenswrapper[4767]: I0317 16:58:09.837358 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562778-5cxf5" podStartSLOduration=8.815806432 podStartE2EDuration="9.837311114s" podCreationTimestamp="2026-03-17 16:58:00 +0000 UTC" firstStartedPulling="2026-03-17 16:58:07.307527904 +0000 UTC m=+4878.720843961" lastFinishedPulling="2026-03-17 16:58:08.329032596 +0000 UTC m=+4879.742348643" observedRunningTime="2026-03-17 16:58:09.816441921 +0000 UTC m=+4881.229757978" watchObservedRunningTime="2026-03-17 16:58:09.837311114 +0000 UTC m=+4881.250627161" Mar 17 16:58:10 crc kubenswrapper[4767]: I0317 16:58:10.814795 4767 generic.go:334] "Generic (PLEG): container finished" podID="f71df388-7598-46d0-82f6-5a6aa85d90ad" containerID="f752ad2aed374732f44f8dca7dc468c6deeb94183f06dfb477226dabf512e86a" exitCode=0 Mar 17 16:58:10 crc kubenswrapper[4767]: I0317 16:58:10.815136 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562778-5cxf5" event={"ID":"f71df388-7598-46d0-82f6-5a6aa85d90ad","Type":"ContainerDied","Data":"f752ad2aed374732f44f8dca7dc468c6deeb94183f06dfb477226dabf512e86a"} Mar 17 16:58:12 crc kubenswrapper[4767]: I0317 16:58:12.310878 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562778-5cxf5" Mar 17 16:58:12 crc kubenswrapper[4767]: I0317 16:58:12.312477 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94jpz\" (UniqueName: \"kubernetes.io/projected/f71df388-7598-46d0-82f6-5a6aa85d90ad-kube-api-access-94jpz\") pod \"f71df388-7598-46d0-82f6-5a6aa85d90ad\" (UID: \"f71df388-7598-46d0-82f6-5a6aa85d90ad\") " Mar 17 16:58:12 crc kubenswrapper[4767]: I0317 16:58:12.322687 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71df388-7598-46d0-82f6-5a6aa85d90ad-kube-api-access-94jpz" (OuterVolumeSpecName: "kube-api-access-94jpz") pod "f71df388-7598-46d0-82f6-5a6aa85d90ad" (UID: "f71df388-7598-46d0-82f6-5a6aa85d90ad"). InnerVolumeSpecName "kube-api-access-94jpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 16:58:12 crc kubenswrapper[4767]: I0317 16:58:12.417869 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94jpz\" (UniqueName: \"kubernetes.io/projected/f71df388-7598-46d0-82f6-5a6aa85d90ad-kube-api-access-94jpz\") on node \"crc\" DevicePath \"\"" Mar 17 16:58:12 crc kubenswrapper[4767]: I0317 16:58:12.473425 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562772-65nvt"] Mar 17 16:58:12 crc kubenswrapper[4767]: I0317 16:58:12.486553 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562772-65nvt"] Mar 17 16:58:12 crc kubenswrapper[4767]: I0317 16:58:12.839551 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562778-5cxf5" event={"ID":"f71df388-7598-46d0-82f6-5a6aa85d90ad","Type":"ContainerDied","Data":"8e0e7008802a8282bf68e1ee980a10c49c11ec86a475e334d6c4cac341e5c918"} Mar 17 16:58:12 crc kubenswrapper[4767]: I0317 16:58:12.839873 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e0e7008802a8282bf68e1ee980a10c49c11ec86a475e334d6c4cac341e5c918" Mar 17 16:58:12 crc kubenswrapper[4767]: I0317 16:58:12.839616 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562778-5cxf5" Mar 17 16:58:13 crc kubenswrapper[4767]: I0317 16:58:13.374232 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dba409ed-88cf-45a7-85b1-8fcc82c8abd8" path="/var/lib/kubelet/pods/dba409ed-88cf-45a7-85b1-8fcc82c8abd8/volumes" Mar 17 16:58:24 crc kubenswrapper[4767]: I0317 16:58:24.881136 4767 scope.go:117] "RemoveContainer" containerID="96895fe5560233d76c92200f58b32efd4a2b869d942298f9a8b441c64bb12e84" Mar 17 16:58:34 crc kubenswrapper[4767]: I0317 16:58:34.167080 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:58:34 crc kubenswrapper[4767]: I0317 16:58:34.167780 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:58:36 crc kubenswrapper[4767]: E0317 16:58:36.447727 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Mar 17 16:58:36 crc kubenswrapper[4767]: E0317 16:58:36.452121 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k4zw7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(909f6b35-823c-4b1b-ba3b-6db1a4f6b444): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 16:58:36 crc kubenswrapper[4767]: E0317 16:58:36.453246 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="909f6b35-823c-4b1b-ba3b-6db1a4f6b444" Mar 17 16:58:37 crc kubenswrapper[4767]: E0317 16:58:37.313051 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="909f6b35-823c-4b1b-ba3b-6db1a4f6b444" Mar 17 16:58:51 crc kubenswrapper[4767]: I0317 16:58:51.897356 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 17 16:58:54 crc kubenswrapper[4767]: I0317 16:58:54.733476 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"909f6b35-823c-4b1b-ba3b-6db1a4f6b444","Type":"ContainerStarted","Data":"c8b30c60aed4f751fdd0bc2ff9a5fc39033ebe3dc1a35d61b2a2e94fd217f757"} Mar 17 16:58:54 crc kubenswrapper[4767]: I0317 16:58:54.760876 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=5.2370263 podStartE2EDuration="1m0.760852279s" podCreationTimestamp="2026-03-17 16:57:54 +0000 UTC" firstStartedPulling="2026-03-17 16:57:56.367376878 +0000 UTC m=+4867.780692925" lastFinishedPulling="2026-03-17 16:58:51.891202857 +0000 UTC m=+4923.304518904" observedRunningTime="2026-03-17 16:58:54.756372622 +0000 UTC m=+4926.169688689" watchObservedRunningTime="2026-03-17 16:58:54.760852279 +0000 UTC m=+4926.174168326" Mar 17 16:59:04 crc kubenswrapper[4767]: I0317 16:59:04.167189 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 16:59:04 crc kubenswrapper[4767]: I0317 16:59:04.168123 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 16:59:04 crc kubenswrapper[4767]: I0317 16:59:04.168217 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 16:59:04 crc kubenswrapper[4767]: I0317 16:59:04.169599 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c4883088f1ca02db880affb212da5ef91b29d53adeb7d812b4a4c5c2a0d946f2"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 16:59:04 crc kubenswrapper[4767]: I0317 16:59:04.169702 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://c4883088f1ca02db880affb212da5ef91b29d53adeb7d812b4a4c5c2a0d946f2" gracePeriod=600 Mar 17 16:59:05 crc kubenswrapper[4767]: I0317 16:59:05.234690 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="c4883088f1ca02db880affb212da5ef91b29d53adeb7d812b4a4c5c2a0d946f2" exitCode=0 Mar 17 16:59:05 crc kubenswrapper[4767]: I0317 16:59:05.234781 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"c4883088f1ca02db880affb212da5ef91b29d53adeb7d812b4a4c5c2a0d946f2"} Mar 17 16:59:05 crc kubenswrapper[4767]: I0317 16:59:05.235056 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e"} Mar 17 16:59:05 crc kubenswrapper[4767]: I0317 16:59:05.235082 4767 scope.go:117] "RemoveContainer" containerID="defcd8dfa801b203cf92d6e975a9ea880ca452d200b74faad61bf003a4121216" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.632864 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562780-s2ljx"] Mar 17 17:00:00 crc kubenswrapper[4767]: E0317 17:00:00.638067 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f71df388-7598-46d0-82f6-5a6aa85d90ad" containerName="oc" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.638118 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f71df388-7598-46d0-82f6-5a6aa85d90ad" containerName="oc" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.640750 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f71df388-7598-46d0-82f6-5a6aa85d90ad" containerName="oc" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.662550 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd"] Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.671385 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562780-s2ljx" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.674772 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.694448 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-secret-volume\") pod \"collect-profiles-29562780-zkgdd\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.694891 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-config-volume\") pod \"collect-profiles-29562780-zkgdd\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.695300 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msc94\" (UniqueName: \"kubernetes.io/projected/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-kube-api-access-msc94\") pod \"collect-profiles-29562780-zkgdd\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.699908 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.699898 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.700098 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.700661 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.703166 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.798552 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msc94\" (UniqueName: \"kubernetes.io/projected/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-kube-api-access-msc94\") pod \"collect-profiles-29562780-zkgdd\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.798772 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-secret-volume\") pod \"collect-profiles-29562780-zkgdd\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.798842 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-config-volume\") pod \"collect-profiles-29562780-zkgdd\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.798901 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlkfl\" (UniqueName: \"kubernetes.io/projected/1af06aea-252b-458b-8180-2ec0bba5f0c2-kube-api-access-nlkfl\") pod \"auto-csr-approver-29562780-s2ljx\" (UID: \"1af06aea-252b-458b-8180-2ec0bba5f0c2\") " pod="openshift-infra/auto-csr-approver-29562780-s2ljx" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.809416 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562780-s2ljx"] Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.811923 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-config-volume\") pod \"collect-profiles-29562780-zkgdd\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.834404 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-secret-volume\") pod \"collect-profiles-29562780-zkgdd\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.835647 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msc94\" (UniqueName: \"kubernetes.io/projected/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-kube-api-access-msc94\") pod \"collect-profiles-29562780-zkgdd\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.844376 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd"] Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.902177 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlkfl\" (UniqueName: \"kubernetes.io/projected/1af06aea-252b-458b-8180-2ec0bba5f0c2-kube-api-access-nlkfl\") pod \"auto-csr-approver-29562780-s2ljx\" (UID: \"1af06aea-252b-458b-8180-2ec0bba5f0c2\") " pod="openshift-infra/auto-csr-approver-29562780-s2ljx" Mar 17 17:00:00 crc kubenswrapper[4767]: I0317 17:00:00.928307 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlkfl\" (UniqueName: \"kubernetes.io/projected/1af06aea-252b-458b-8180-2ec0bba5f0c2-kube-api-access-nlkfl\") pod \"auto-csr-approver-29562780-s2ljx\" (UID: \"1af06aea-252b-458b-8180-2ec0bba5f0c2\") " pod="openshift-infra/auto-csr-approver-29562780-s2ljx" Mar 17 17:00:01 crc kubenswrapper[4767]: I0317 17:00:01.035154 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:01 crc kubenswrapper[4767]: I0317 17:00:01.037215 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562780-s2ljx" Mar 17 17:00:02 crc kubenswrapper[4767]: I0317 17:00:02.736761 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562780-s2ljx"] Mar 17 17:00:02 crc kubenswrapper[4767]: I0317 17:00:02.756414 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd"] Mar 17 17:00:03 crc kubenswrapper[4767]: I0317 17:00:03.642942 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562780-s2ljx" event={"ID":"1af06aea-252b-458b-8180-2ec0bba5f0c2","Type":"ContainerStarted","Data":"c3a6064f2a2f1d16b05e6d6e6e2119c066d4f387821ee123a841f8fc547fe2d6"} Mar 17 17:00:03 crc kubenswrapper[4767]: I0317 17:00:03.644877 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" event={"ID":"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa","Type":"ContainerStarted","Data":"fff19e8917247e305264a6e028ec8ed293688c2d6a8888c8ac08c909c291bcbf"} Mar 17 17:00:03 crc kubenswrapper[4767]: I0317 17:00:03.645000 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" event={"ID":"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa","Type":"ContainerStarted","Data":"b3cba023c2bfd0fab508279e7421324b6f1a28feb42860d62d63424729da42b6"} Mar 17 17:00:03 crc kubenswrapper[4767]: I0317 17:00:03.671699 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" podStartSLOduration=3.670140839 podStartE2EDuration="3.670140839s" podCreationTimestamp="2026-03-17 17:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 17:00:03.663391197 +0000 UTC m=+4995.076707264" watchObservedRunningTime="2026-03-17 17:00:03.670140839 +0000 UTC m=+4995.083456876" Mar 17 17:00:04 crc kubenswrapper[4767]: I0317 17:00:04.665043 4767 generic.go:334] "Generic (PLEG): container finished" podID="1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa" containerID="fff19e8917247e305264a6e028ec8ed293688c2d6a8888c8ac08c909c291bcbf" exitCode=0 Mar 17 17:00:04 crc kubenswrapper[4767]: I0317 17:00:04.665464 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" event={"ID":"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa","Type":"ContainerDied","Data":"fff19e8917247e305264a6e028ec8ed293688c2d6a8888c8ac08c909c291bcbf"} Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.259775 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.454097 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-secret-volume\") pod \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.454752 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-config-volume\") pod \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.455071 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msc94\" (UniqueName: \"kubernetes.io/projected/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-kube-api-access-msc94\") pod \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\" (UID: \"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa\") " Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.469294 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-config-volume" (OuterVolumeSpecName: "config-volume") pod "1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa" (UID: "1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.475455 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-kube-api-access-msc94" (OuterVolumeSpecName: "kube-api-access-msc94") pod "1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa" (UID: "1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa"). InnerVolumeSpecName "kube-api-access-msc94". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.476534 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa" (UID: "1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.565859 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.565903 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.565914 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msc94\" (UniqueName: \"kubernetes.io/projected/1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa-kube-api-access-msc94\") on node \"crc\" DevicePath \"\"" Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.692205 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" event={"ID":"1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa","Type":"ContainerDied","Data":"b3cba023c2bfd0fab508279e7421324b6f1a28feb42860d62d63424729da42b6"} Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.692252 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3cba023c2bfd0fab508279e7421324b6f1a28feb42860d62d63424729da42b6" Mar 17 17:00:06 crc kubenswrapper[4767]: I0317 17:00:06.692317 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562780-zkgdd" Mar 17 17:00:07 crc kubenswrapper[4767]: I0317 17:00:07.539875 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl"] Mar 17 17:00:07 crc kubenswrapper[4767]: I0317 17:00:07.601822 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562735-z7fsl"] Mar 17 17:00:07 crc kubenswrapper[4767]: I0317 17:00:07.708522 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562780-s2ljx" event={"ID":"1af06aea-252b-458b-8180-2ec0bba5f0c2","Type":"ContainerStarted","Data":"f0dab67d7e1a35ec5caea2c1b2615e5b4583111aac4e3ad25d3f29e990a838ee"} Mar 17 17:00:07 crc kubenswrapper[4767]: I0317 17:00:07.742445 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562780-s2ljx" podStartSLOduration=5.075965246 podStartE2EDuration="7.742411017s" podCreationTimestamp="2026-03-17 17:00:00 +0000 UTC" firstStartedPulling="2026-03-17 17:00:02.783158567 +0000 UTC m=+4994.196474614" lastFinishedPulling="2026-03-17 17:00:05.449604338 +0000 UTC m=+4996.862920385" observedRunningTime="2026-03-17 17:00:07.726635319 +0000 UTC m=+4999.139951376" watchObservedRunningTime="2026-03-17 17:00:07.742411017 +0000 UTC m=+4999.155727064" Mar 17 17:00:09 crc kubenswrapper[4767]: I0317 17:00:09.381317 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ab4af60-5262-4a85-b749-961ca0b7b162" path="/var/lib/kubelet/pods/7ab4af60-5262-4a85-b749-961ca0b7b162/volumes" Mar 17 17:00:09 crc kubenswrapper[4767]: I0317 17:00:09.877229 4767 generic.go:334] "Generic (PLEG): container finished" podID="1af06aea-252b-458b-8180-2ec0bba5f0c2" containerID="f0dab67d7e1a35ec5caea2c1b2615e5b4583111aac4e3ad25d3f29e990a838ee" exitCode=0 Mar 17 17:00:09 crc kubenswrapper[4767]: I0317 17:00:09.877288 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562780-s2ljx" event={"ID":"1af06aea-252b-458b-8180-2ec0bba5f0c2","Type":"ContainerDied","Data":"f0dab67d7e1a35ec5caea2c1b2615e5b4583111aac4e3ad25d3f29e990a838ee"} Mar 17 17:00:12 crc kubenswrapper[4767]: I0317 17:00:12.154214 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562780-s2ljx" Mar 17 17:00:12 crc kubenswrapper[4767]: I0317 17:00:12.234482 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlkfl\" (UniqueName: \"kubernetes.io/projected/1af06aea-252b-458b-8180-2ec0bba5f0c2-kube-api-access-nlkfl\") pod \"1af06aea-252b-458b-8180-2ec0bba5f0c2\" (UID: \"1af06aea-252b-458b-8180-2ec0bba5f0c2\") " Mar 17 17:00:12 crc kubenswrapper[4767]: I0317 17:00:12.291555 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af06aea-252b-458b-8180-2ec0bba5f0c2-kube-api-access-nlkfl" (OuterVolumeSpecName: "kube-api-access-nlkfl") pod "1af06aea-252b-458b-8180-2ec0bba5f0c2" (UID: "1af06aea-252b-458b-8180-2ec0bba5f0c2"). InnerVolumeSpecName "kube-api-access-nlkfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 17:00:12 crc kubenswrapper[4767]: I0317 17:00:12.339260 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlkfl\" (UniqueName: \"kubernetes.io/projected/1af06aea-252b-458b-8180-2ec0bba5f0c2-kube-api-access-nlkfl\") on node \"crc\" DevicePath \"\"" Mar 17 17:00:12 crc kubenswrapper[4767]: I0317 17:00:12.934992 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562780-s2ljx" event={"ID":"1af06aea-252b-458b-8180-2ec0bba5f0c2","Type":"ContainerDied","Data":"c3a6064f2a2f1d16b05e6d6e6e2119c066d4f387821ee123a841f8fc547fe2d6"} Mar 17 17:00:12 crc kubenswrapper[4767]: I0317 17:00:12.935057 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3a6064f2a2f1d16b05e6d6e6e2119c066d4f387821ee123a841f8fc547fe2d6" Mar 17 17:00:12 crc kubenswrapper[4767]: I0317 17:00:12.935087 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562780-s2ljx" Mar 17 17:00:13 crc kubenswrapper[4767]: I0317 17:00:13.370975 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562774-57plp"] Mar 17 17:00:13 crc kubenswrapper[4767]: I0317 17:00:13.377294 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562774-57plp"] Mar 17 17:00:15 crc kubenswrapper[4767]: I0317 17:00:15.378703 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f42657e8-7838-487f-82f2-f36dc5340922" path="/var/lib/kubelet/pods/f42657e8-7838-487f-82f2-f36dc5340922/volumes" Mar 17 17:00:36 crc kubenswrapper[4767]: I0317 17:00:36.769741 4767 scope.go:117] "RemoveContainer" containerID="881f1867e56ab09c9385353926ad2b79d8af28c1cae6ef12f5934ce58d1608b2" Mar 17 17:00:36 crc kubenswrapper[4767]: I0317 17:00:36.927014 4767 scope.go:117] "RemoveContainer" containerID="fb8a85a4359e575e2a1e1f8705de9ebf2407ff02271eb7a595a0f10b177b76d0" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.480476 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29562781-d47zt"] Mar 17 17:01:01 crc kubenswrapper[4767]: E0317 17:01:01.488492 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa" containerName="collect-profiles" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.488961 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa" containerName="collect-profiles" Mar 17 17:01:01 crc kubenswrapper[4767]: E0317 17:01:01.490354 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af06aea-252b-458b-8180-2ec0bba5f0c2" containerName="oc" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.490375 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af06aea-252b-458b-8180-2ec0bba5f0c2" containerName="oc" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.493026 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ff0dbd9-58e1-42d7-8d35-c71efa9b00fa" containerName="collect-profiles" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.493095 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af06aea-252b-458b-8180-2ec0bba5f0c2" containerName="oc" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.502159 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.592274 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-config-data\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.592997 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-combined-ca-bundle\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.593231 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-fernet-keys\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.593488 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h9fc\" (UniqueName: \"kubernetes.io/projected/d38f4981-a828-4985-9162-9597fd4dfc6d-kube-api-access-9h9fc\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.699443 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-combined-ca-bundle\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.699744 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-fernet-keys\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.700339 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h9fc\" (UniqueName: \"kubernetes.io/projected/d38f4981-a828-4985-9162-9597fd4dfc6d-kube-api-access-9h9fc\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.700484 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-config-data\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.767662 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-fernet-keys\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.768054 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-config-data\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.769460 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-combined-ca-bundle\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.770947 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h9fc\" (UniqueName: \"kubernetes.io/projected/d38f4981-a828-4985-9162-9597fd4dfc6d-kube-api-access-9h9fc\") pod \"keystone-cron-29562781-d47zt\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.905581 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:01:01 crc kubenswrapper[4767]: I0317 17:01:01.951982 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29562781-d47zt"] Mar 17 17:01:03 crc kubenswrapper[4767]: I0317 17:01:03.745549 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-wj8xp" podUID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:03 crc kubenswrapper[4767]: I0317 17:01:03.749410 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:03 crc kubenswrapper[4767]: I0317 17:01:03.749549 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:04 crc kubenswrapper[4767]: I0317 17:01:04.166512 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 17:01:04 crc kubenswrapper[4767]: I0317 17:01:04.168924 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 17:01:06 crc kubenswrapper[4767]: I0317 17:01:06.134451 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:06 crc kubenswrapper[4767]: I0317 17:01:06.136324 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:06 crc kubenswrapper[4767]: I0317 17:01:06.166472 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29562781-d47zt"] Mar 17 17:01:06 crc kubenswrapper[4767]: I0317 17:01:06.483542 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" podUID="faf77dfd-c76e-4abd-a8c8-b6211cf070a4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.125:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:06 crc kubenswrapper[4767]: I0317 17:01:06.969565 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29562781-d47zt" event={"ID":"d38f4981-a828-4985-9162-9597fd4dfc6d","Type":"ContainerStarted","Data":"eb4e7a2dd2b9933ec39645e7e0f00e2d45b522e1e2f64e225e3e74274c1fefe6"} Mar 17 17:01:07 crc kubenswrapper[4767]: I0317 17:01:07.206614 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-wjz85" podUID="1cfcd066-3d8a-431e-a895-a257a3e1baef" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:07 crc kubenswrapper[4767]: I0317 17:01:07.207211 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-wjz85" podUID="1cfcd066-3d8a-431e-a895-a257a3e1baef" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:08 crc kubenswrapper[4767]: I0317 17:01:08.029346 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:08 crc kubenswrapper[4767]: I0317 17:01:08.029346 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:08 crc kubenswrapper[4767]: I0317 17:01:08.030717 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:08 crc kubenswrapper[4767]: I0317 17:01:08.030782 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:08 crc kubenswrapper[4767]: I0317 17:01:08.038890 4767 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-mdlms container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:08 crc kubenswrapper[4767]: I0317 17:01:08.039020 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" podUID="a8829d5d-100e-4518-8863-db3ab9c7b30d" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:08 crc kubenswrapper[4767]: I0317 17:01:08.831799 4767 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-82k6b container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:08 crc kubenswrapper[4767]: I0317 17:01:08.832203 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:08 crc kubenswrapper[4767]: I0317 17:01:08.831850 4767 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-82k6b container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:08 crc kubenswrapper[4767]: I0317 17:01:08.832372 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:09 crc kubenswrapper[4767]: I0317 17:01:09.041614 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29562781-d47zt" event={"ID":"d38f4981-a828-4985-9162-9597fd4dfc6d","Type":"ContainerStarted","Data":"e52c8a922ac5726ae258ab86cf09377d32f8e065ff7f13bc06511afd70ebdfb0"} Mar 17 17:01:09 crc kubenswrapper[4767]: I0317 17:01:09.950504 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" podUID="95d5d3e8-dc72-414f-afe4-b68b757a39d4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:09 crc kubenswrapper[4767]: I0317 17:01:09.950529 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" podUID="95d5d3e8-dc72-414f-afe4-b68b757a39d4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.348453 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.348479 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.349080 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.349213 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.539497 4767 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-lzjm6 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.539641 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" podUID="68f6ff1d-a119-4faf-89aa-fd9afc2d93f8" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580379 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580430 4767 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-lzjm6 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580451 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580509 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580512 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" podUID="68f6ff1d-a119-4faf-89aa-fd9afc2d93f8" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580606 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580685 4767 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580709 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580756 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580775 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580801 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:10 crc kubenswrapper[4767]: I0317 17:01:10.580814 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.114477 4767 patch_prober.go:28] interesting pod/controller-manager-57bc79d78-25d96 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.114591 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podUID="1767bf72-c95d-486d-baa6-ca741742569b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.114708 4767 patch_prober.go:28] interesting pod/route-controller-manager-56d7bf8c9b-cs8b8 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.114728 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" podUID="a9f5b566-a215-4322-8d74-668dc2a93f3e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.155482 4767 patch_prober.go:28] interesting pod/route-controller-manager-56d7bf8c9b-cs8b8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.155615 4767 patch_prober.go:28] interesting pod/controller-manager-57bc79d78-25d96 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.155650 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" podUID="a9f5b566-a215-4322-8d74-668dc2a93f3e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.155691 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podUID="1767bf72-c95d-486d-baa6-ca741742569b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.155776 4767 patch_prober.go:28] interesting pod/oauth-openshift-5cc5b65bd-6nll7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.155861 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" podUID="2321f9fb-78c8-401b-8fc2-a5f82d42a226" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.155947 4767 patch_prober.go:28] interesting pod/oauth-openshift-5cc5b65bd-6nll7 container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.155963 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" podUID="2321f9fb-78c8-401b-8fc2-a5f82d42a226" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.156112 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.156181 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.156227 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.156294 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-8q5h4 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.156296 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.156323 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.156355 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.56:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.156783 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podUID="560edfa4-8a88-4c9b-8b31-e61f93050c15" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.535436 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" podUID="70fbe8aa-8647-4fe8-914c-0c05399cf46d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.535783 4767 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.535815 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.535908 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" podUID="70fbe8aa-8647-4fe8-914c-0c05399cf46d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.620490 4767 patch_prober.go:28] interesting pod/thanos-querier-89d788fd9-2skkr container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.82:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.620909 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" podUID="8ab6a6b4-cb73-4701-a29c-f404c58966c4" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.82:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.620616 4767 patch_prober.go:28] interesting pod/thanos-querier-89d788fd9-2skkr container/kube-rbac-proxy-web namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.82:9091/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:11 crc kubenswrapper[4767]: I0317 17:01:11.621060 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" podUID="8ab6a6b4-cb73-4701-a29c-f404c58966c4" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.82:9091/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:12 crc kubenswrapper[4767]: I0317 17:01:12.152561 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" podUID="713154c5-7e16-498f-9612-1c0afbf362bb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:12 crc kubenswrapper[4767]: I0317 17:01:12.152591 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" podUID="713154c5-7e16-498f-9612-1c0afbf362bb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:12 crc kubenswrapper[4767]: I0317 17:01:12.197524 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" podUID="877f4eda-1ec7-4296-98df-b5ca7a7fa78a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.96:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:12 crc kubenswrapper[4767]: I0317 17:01:12.378786 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="94b1e014-c27b-4b81-a12d-eecfe02d5b67" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:12 crc kubenswrapper[4767]: I0317 17:01:12.378923 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="94b1e014-c27b-4b81-a12d-eecfe02d5b67" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:12 crc kubenswrapper[4767]: I0317 17:01:12.562420 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" podUID="fda1c393-424a-4142-8570-5dc108f6d6f4" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:12 crc kubenswrapper[4767]: I0317 17:01:12.562453 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" podUID="fda1c393-424a-4142-8570-5dc108f6d6f4" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.396484 4767 patch_prober.go:28] interesting pod/console-89866dfb6-fswnc container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.146:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.396789 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-89866dfb6-fswnc" podUID="b52b7ad3-98cb-4051-8e3b-665fa44b0fd6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.146:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.543348 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-wj8xp" podUID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.760548 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="d24c721d-4968-477f-ba8e-23f6d31629d2" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.760548 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="d24c721d-4968-477f-ba8e-23f6d31629d2" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.806963 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-wj8xp" podUID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.807010 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-wj8xp" podUID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.806978 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" podUID="07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.810116 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.810838 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.858313 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" podUID="07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:13 crc kubenswrapper[4767]: I0317 17:01:13.858429 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" podUID="baf4c43e-08e2-4afa-b9b5-8bb4aa8f0352" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.076534 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" podUID="53aea289-5a28-438b-8d28-242d836351f0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.076982 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" podUID="b6dc3b02-8dc0-46ac-b78e-1d7d3ffb3f24" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.077034 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" podUID="53aea289-5a28-438b-8d28-242d836351f0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.077082 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" podUID="b6dc3b02-8dc0-46ac-b78e-1d7d3ffb3f24" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.237126 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-7bb4cc7c98-nchz8" podUID="b978bf52-fdb5-4863-a5bf-e2aec8ab08a4" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.99:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.237194 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-7bb4cc7c98-nchz8" podUID="b978bf52-fdb5-4863-a5bf-e2aec8ab08a4" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.99:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.387447 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" podUID="a67a8ddc-0801-4530-bf82-9d4f4d0389a8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.387477 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" podUID="a67a8ddc-0801-4530-bf82-9d4f4d0389a8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.443496 4767 patch_prober.go:28] interesting pod/nmstate-webhook-5f558f5558-rwq8w container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.89:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.443577 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" podUID="31f30608-3b0d-4f63-9ab8-6a1547e233ff" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.89:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.484448 4767 patch_prober.go:28] interesting pod/loki-operator-controller-manager-5dcc96cb6-rv7xm container/manager namespace/openshift-operators-redhat: Liveness probe status=failure output="Get \"http://10.217.0.50:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.484571 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" podUID="6f682203-ee25-4b31-957b-26d8148feda0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.50:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.566388 4767 patch_prober.go:28] interesting pod/logging-loki-distributor-9c6b6d984-q9dnr container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.566470 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" podUID="28ac74bb-3f68-478d-8f50-acf61b2e8223" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.53:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.566787 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" podUID="0163654c-d57e-4b14-aba0-f76dbaff1114" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.648475 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" podUID="0163654c-d57e-4b14-aba0-f76dbaff1114" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.648544 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" podUID="fa1e1603-adc4-4ade-aca8-e59df8c350b2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.702275 4767 trace.go:236] Trace[1467219533]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-j7tgx" (17-Mar-2026 17:01:10.388) (total time: 4301ms): Mar 17 17:01:14 crc kubenswrapper[4767]: Trace[1467219533]: [4.301058956s] [4.301058956s] END Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.731686 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" podUID="904d9f3b-95f0-4e57-8d04-c2026227a4a6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.732338 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" podUID="904d9f3b-95f0-4e57-8d04-c2026227a4a6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.731790 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" podUID="fa1e1603-adc4-4ade-aca8-e59df8c350b2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.732501 4767 patch_prober.go:28] interesting pod/logging-loki-querier-6dcbdf8bb8-679nd container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.732544 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" podUID="66451b46-d0f5-4037-aa4d-abf0143b2ef8" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.750269 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="6abe028c-416e-4978-bcc8-3a7b8d92624b" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.750307 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="60ee8af2-4ae0-402e-8a13-2dce36db4bab" containerName="prometheus" probeResult="failure" output="command timed out" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.750880 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="6abe028c-416e-4978-bcc8-3a7b8d92624b" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.751392 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-6jnww" podUID="d4418b02-160e-4b8a-8232-e194eb5ac0f3" containerName="nmstate-handler" probeResult="failure" output="command timed out" Mar 17 17:01:14 crc kubenswrapper[4767]: I0317 17:01:14.752559 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="60ee8af2-4ae0-402e-8a13-2dce36db4bab" containerName="prometheus" probeResult="failure" output="command timed out" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.074083 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" podUID="a7b5a1f5-da92-46dd-a5b0-5088e75346af" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.116611 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" podUID="a7b5a1f5-da92-46dd-a5b0-5088e75346af" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.280463 4767 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vl8vb container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.69:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.280440 4767 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vl8vb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.69:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.280566 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" podUID="eafb476c-0d04-4608-b1e3-cf8b6ebb74bf" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.69:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.280649 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vl8vb" podUID="eafb476c-0d04-4608-b1e3-cf8b6ebb74bf" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.69:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.387098 4767 patch_prober.go:28] interesting pod/logging-loki-query-frontend-ff66c4dc9-4jtkt container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.387201 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" podUID="ce1ebb9b-1558-4190-9e47-4ef6ba3160da" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.496463 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" podUID="43b467d2-9860-4feb-a656-65827836c23e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.496648 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" podUID="43b467d2-9860-4feb-a656-65827836c23e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.503702 4767 patch_prober.go:28] interesting pod/metrics-server-56684ddd8d-zrdxq container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.84:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.503799 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" podUID="5a2cfb18-0930-434f-a9bc-824575209dda" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.84:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.716453 4767 patch_prober.go:28] interesting pod/observability-operator-6dd7dd855f-hw4fk container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.716540 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" podUID="bde9e722-7f00-444e-92fe-ac0db62b6172" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.716631 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" podUID="921c5e77-9858-4177-99a0-8cd3a4420d7d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.751967 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output="command timed out" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.754868 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output="command timed out" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.799363 4767 patch_prober.go:28] interesting pod/observability-operator-6dd7dd855f-hw4fk container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.799367 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" podUID="a833dc16-3a29-4129-b592-732d71818bc4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.799599 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" podUID="921c5e77-9858-4177-99a0-8cd3a4420d7d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.799799 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" podUID="bde9e722-7f00-444e-92fe-ac0db62b6172" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.801235 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-sl4m8" podUID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:15 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:15 crc kubenswrapper[4767]: > Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.882672 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" podUID="bd2dbd72-69bf-40c1-b591-be3782c33465" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.882921 4767 patch_prober.go:28] interesting pod/monitoring-plugin-6bbd5f6758-kwcw5 container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.85:9443/health\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.882957 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" podUID="93033ac4-fd81-46eb-8014-184056ab6de2" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.85:9443/health\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.883001 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" podUID="a833dc16-3a29-4129-b592-732d71818bc4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.965431 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" podUID="bd2dbd72-69bf-40c1-b591-be3782c33465" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:15 crc kubenswrapper[4767]: I0317 17:01:15.965469 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" podUID="59ba3b25-1e95-41ad-921c-9ee4ec5e2c43" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.047762 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" podUID="59ba3b25-1e95-41ad-921c-9ee4ec5e2c43" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.047873 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" podUID="7381ff68-5e5d-4281-9924-8495ece760f4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.048236 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" podUID="e59ff072-51c2-4995-9f61-709f12a1393a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.048279 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" podUID="e59ff072-51c2-4995-9f61-709f12a1393a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.093657 4767 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.093749 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="89ce8746-1721-4dec-9742-0abc2f498de3" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.123604 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.123725 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.123881 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.123936 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.56:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.136952 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-8q5h4 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.137004 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-8q5h4 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.137042 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podUID="560edfa4-8a88-4c9b-8b31-e61f93050c15" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.137085 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podUID="560edfa4-8a88-4c9b-8b31-e61f93050c15" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.154875 4767 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.154986 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a2389106-04cc-4871-b544-f8a677a6405a" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.58:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.166466 4767 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.166556 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="27c143b3-0f5b-4265-a2e2-d4302a2f3c70" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.211896 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-sl4m8" podUID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:16 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:16 crc kubenswrapper[4767]: > Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.329360 4767 patch_prober.go:28] interesting pod/perses-operator-6d5d56fd99-rxwwn container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.23:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.329431 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" podUID="2531b56b-9f38-4cef-9812-b1b30c29405a" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.23:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.329501 4767 patch_prober.go:28] interesting pod/perses-operator-6d5d56fd99-rxwwn container/perses-operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.23:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.329515 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" podUID="2531b56b-9f38-4cef-9812-b1b30c29405a" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.23:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.524684 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" podUID="faf77dfd-c76e-4abd-a8c8-b6211cf070a4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.125:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.524803 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" podUID="faf77dfd-c76e-4abd-a8c8-b6211cf070a4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.125:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.598558 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-5tr6k" podUID="cc1f08cb-7205-4537-acca-91edee17aef9" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:16 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:16 crc kubenswrapper[4767]: > Mar 17 17:01:16 crc kubenswrapper[4767]: I0317 17:01:16.606381 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-5tr6k" podUID="cc1f08cb-7205-4537-acca-91edee17aef9" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:16 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:16 crc kubenswrapper[4767]: > Mar 17 17:01:17 crc kubenswrapper[4767]: I0317 17:01:17.198426 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-wjz85" podUID="1cfcd066-3d8a-431e-a895-a257a3e1baef" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:17 crc kubenswrapper[4767]: I0317 17:01:17.198435 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-wjz85" podUID="1cfcd066-3d8a-431e-a895-a257a3e1baef" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:17 crc kubenswrapper[4767]: I0317 17:01:17.895947 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:17 crc kubenswrapper[4767]: I0317 17:01:17.900892 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:17 crc kubenswrapper[4767]: I0317 17:01:17.899428 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:17 crc kubenswrapper[4767]: I0317 17:01:17.901004 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:18 crc kubenswrapper[4767]: I0317 17:01:18.018360 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29562781-d47zt" podStartSLOduration=18.014549912 podStartE2EDuration="18.014549912s" podCreationTimestamp="2026-03-17 17:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 17:01:17.909506439 +0000 UTC m=+5069.322822496" watchObservedRunningTime="2026-03-17 17:01:18.014549912 +0000 UTC m=+5069.427865949" Mar 17 17:01:18 crc kubenswrapper[4767]: I0317 17:01:18.037937 4767 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-mdlms container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": context deadline exceeded" start-of-body= Mar 17 17:01:18 crc kubenswrapper[4767]: I0317 17:01:18.038026 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" podUID="a8829d5d-100e-4518-8863-db3ab9c7b30d" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": context deadline exceeded" Mar 17 17:01:18 crc kubenswrapper[4767]: I0317 17:01:18.756840 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-mdghs" podUID="571d124c-f5cb-4c5e-bfd7-3f56836ce800" containerName="registry-server" probeResult="failure" output="command timed out" Mar 17 17:01:18 crc kubenswrapper[4767]: I0317 17:01:18.757949 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-mdghs" podUID="571d124c-f5cb-4c5e-bfd7-3f56836ce800" containerName="registry-server" probeResult="failure" output="command timed out" Mar 17 17:01:18 crc kubenswrapper[4767]: I0317 17:01:18.831221 4767 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-82k6b container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:18 crc kubenswrapper[4767]: I0317 17:01:18.831287 4767 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-82k6b container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.76:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:18 crc kubenswrapper[4767]: I0317 17:01:18.831377 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.76:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:18 crc kubenswrapper[4767]: I0317 17:01:18.831310 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:19 crc kubenswrapper[4767]: I0317 17:01:19.101460 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-hqqbp" podUID="349fc544-d564-4b99-bac0-bfba49fe72db" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:19 crc kubenswrapper[4767]: timeout: health rpc did not complete within 1s Mar 17 17:01:19 crc kubenswrapper[4767]: > Mar 17 17:01:19 crc kubenswrapper[4767]: I0317 17:01:19.183002 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-hqqbp" podUID="349fc544-d564-4b99-bac0-bfba49fe72db" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:19 crc kubenswrapper[4767]: timeout: health rpc did not complete within 1s Mar 17 17:01:19 crc kubenswrapper[4767]: > Mar 17 17:01:19 crc kubenswrapper[4767]: I0317 17:01:19.354466 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:19 crc kubenswrapper[4767]: I0317 17:01:19.354981 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:19 crc kubenswrapper[4767]: I0317 17:01:19.354526 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:19 crc kubenswrapper[4767]: I0317 17:01:19.355118 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:19 crc kubenswrapper[4767]: I0317 17:01:19.749524 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="60ee8af2-4ae0-402e-8a13-2dce36db4bab" containerName="prometheus" probeResult="failure" output="command timed out" Mar 17 17:01:19 crc kubenswrapper[4767]: I0317 17:01:19.749784 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="60ee8af2-4ae0-402e-8a13-2dce36db4bab" containerName="prometheus" probeResult="failure" output="command timed out" Mar 17 17:01:19 crc kubenswrapper[4767]: I0317 17:01:19.753575 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="d24c721d-4968-477f-ba8e-23f6d31629d2" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Mar 17 17:01:19 crc kubenswrapper[4767]: I0317 17:01:19.907486 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" podUID="95d5d3e8-dc72-414f-afe4-b68b757a39d4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.346586 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.346637 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.346687 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.346720 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.584279 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.584369 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.584721 4767 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-lzjm6 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.584750 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" podUID="68f6ff1d-a119-4faf-89aa-fd9afc2d93f8" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.584791 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.584806 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.584837 4767 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-lzjm6 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.584850 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" podUID="68f6ff1d-a119-4faf-89aa-fd9afc2d93f8" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.584886 4767 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.584898 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.585749 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.585774 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.587963 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:20 crc kubenswrapper[4767]: I0317 17:01:20.588127 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.058598 4767 patch_prober.go:28] interesting pod/controller-manager-57bc79d78-25d96 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.058684 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podUID="1767bf72-c95d-486d-baa6-ca741742569b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.058705 4767 patch_prober.go:28] interesting pod/controller-manager-57bc79d78-25d96 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.058833 4767 patch_prober.go:28] interesting pod/route-controller-manager-56d7bf8c9b-cs8b8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.058830 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podUID="1767bf72-c95d-486d-baa6-ca741742569b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.058855 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" podUID="a9f5b566-a215-4322-8d74-668dc2a93f3e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.058824 4767 patch_prober.go:28] interesting pod/route-controller-manager-56d7bf8c9b-cs8b8 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.058903 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" podUID="a9f5b566-a215-4322-8d74-668dc2a93f3e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.058932 4767 patch_prober.go:28] interesting pod/oauth-openshift-5cc5b65bd-6nll7 container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.058972 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" podUID="2321f9fb-78c8-401b-8fc2-a5f82d42a226" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.059012 4767 patch_prober.go:28] interesting pod/oauth-openshift-5cc5b65bd-6nll7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.059104 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" podUID="2321f9fb-78c8-401b-8fc2-a5f82d42a226" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.118416 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.118471 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.118502 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.118551 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.124074 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.124235 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.137378 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-8q5h4 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.137496 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podUID="560edfa4-8a88-4c9b-8b31-e61f93050c15" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.493847 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" podUID="70fbe8aa-8647-4fe8-914c-0c05399cf46d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.494421 4767 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.494510 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.513228 4767 trace.go:236] Trace[1348606480]: "Calculate volume metrics of ovndbcluster-nb-etc-ovn for pod openstack/ovsdbserver-nb-0" (17-Mar-2026 17:01:15.497) (total time: 6014ms): Mar 17 17:01:21 crc kubenswrapper[4767]: Trace[1348606480]: [6.014261764s] [6.014261764s] END Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.513230 4767 trace.go:236] Trace[787944233]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-cell1-server-0" (17-Mar-2026 17:01:16.244) (total time: 5267ms): Mar 17 17:01:21 crc kubenswrapper[4767]: Trace[787944233]: [5.267641799s] [5.267641799s] END Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.620099 4767 patch_prober.go:28] interesting pod/thanos-querier-89d788fd9-2skkr container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.82:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:21 crc kubenswrapper[4767]: I0317 17:01:21.620312 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" podUID="8ab6a6b4-cb73-4701-a29c-f404c58966c4" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.82:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:22 crc kubenswrapper[4767]: I0317 17:01:22.110435 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" podUID="713154c5-7e16-498f-9612-1c0afbf362bb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:22 crc kubenswrapper[4767]: I0317 17:01:22.197596 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" podUID="877f4eda-1ec7-4296-98df-b5ca7a7fa78a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.96:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:22 crc kubenswrapper[4767]: I0317 17:01:22.378711 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="94b1e014-c27b-4b81-a12d-eecfe02d5b67" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:22 crc kubenswrapper[4767]: I0317 17:01:22.378846 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="94b1e014-c27b-4b81-a12d-eecfe02d5b67" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/healthy\": context deadline exceeded" Mar 17 17:01:22 crc kubenswrapper[4767]: I0317 17:01:22.562649 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" podUID="fda1c393-424a-4142-8570-5dc108f6d6f4" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:22 crc kubenswrapper[4767]: I0317 17:01:22.562693 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" podUID="fda1c393-424a-4142-8570-5dc108f6d6f4" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.400629 4767 patch_prober.go:28] interesting pod/console-89866dfb6-fswnc container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.146:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.405502 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-89866dfb6-fswnc" podUID="b52b7ad3-98cb-4051-8e3b-665fa44b0fd6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.146:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.625465 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-wj8xp" podUID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.625500 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" podUID="07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.625927 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-wj8xp" podUID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.626502 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-wj8xp" podUID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.626882 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" podUID="07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.628025 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/frr-k8s-wj8xp" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.642420 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="frr" containerStatusID={"Type":"cri-o","ID":"fbfb64eda9a1857b0e600f098e323af2c36d82414c8ce1e3bffb3520a89ba819"} pod="metallb-system/frr-k8s-wj8xp" containerMessage="Container frr failed liveness probe, will be restarted" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.643652 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-wj8xp" podUID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerName="frr" containerID="cri-o://fbfb64eda9a1857b0e600f098e323af2c36d82414c8ce1e3bffb3520a89ba819" gracePeriod=2 Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.718443 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" podUID="3ee9406d-d284-417c-83e2-942103e449eb" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.718829 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" podUID="3ee9406d-d284-417c-83e2-942103e449eb" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.103:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.751919 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.752037 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-galera-0" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.752666 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.753440 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac"} pod="openstack/openstack-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.755578 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 17 17:01:23 crc kubenswrapper[4767]: I0317 17:01:23.794313 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" podUID="baf4c43e-08e2-4afa-b9b5-8bb4aa8f0352" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.042650 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-k8995" podUID="b6dc3b02-8dc0-46ac-b78e-1d7d3ffb3f24" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.042782 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" podUID="53aea289-5a28-438b-8d28-242d836351f0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.056086 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.056529 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.056299 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.056913 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.346468 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" podUID="a67a8ddc-0801-4530-bf82-9d4f4d0389a8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.422504 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-p2skw" podUID="16e4d9a1-285b-4221-8a99-55d515bc3356" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.422631 4767 patch_prober.go:28] interesting pod/nmstate-webhook-5f558f5558-rwq8w container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.89:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.422656 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" podUID="31f30608-3b0d-4f63-9ab8-6a1547e233ff" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.89:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.489390 4767 patch_prober.go:28] interesting pod/logging-loki-distributor-9c6b6d984-q9dnr container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.489504 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" podUID="28ac74bb-3f68-478d-8f50-acf61b2e8223" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.53:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.543459 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wj49c" podUID="fa1e1603-adc4-4ade-aca8-e59df8c350b2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.547301 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="bdc5b680-771a-4719-81d6-08efe998c1de" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.19:8081/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.550485 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="bdc5b680-771a-4719-81d6-08efe998c1de" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.19:8080/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.666698 4767 patch_prober.go:28] interesting pod/loki-operator-controller-manager-5dcc96cb6-rv7xm container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.50:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.666699 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" podUID="904d9f3b-95f0-4e57-8d04-c2026227a4a6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.666710 4767 patch_prober.go:28] interesting pod/logging-loki-querier-6dcbdf8bb8-679nd container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.667131 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" podUID="6f682203-ee25-4b31-957b-26d8148feda0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.50:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.667141 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" podUID="66451b46-d0f5-4037-aa4d-abf0143b2ef8" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.787750 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="6abe028c-416e-4978-bcc8-3a7b8d92624b" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.787956 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="60ee8af2-4ae0-402e-8a13-2dce36db4bab" containerName="prometheus" probeResult="failure" output="command timed out" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.787841 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="6abe028c-416e-4978-bcc8-3a7b8d92624b" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.787859 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="60ee8af2-4ae0-402e-8a13-2dce36db4bab" containerName="prometheus" probeResult="failure" output="command timed out" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.787954 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-6jnww" podUID="d4418b02-160e-4b8a-8232-e194eb5ac0f3" containerName="nmstate-handler" probeResult="failure" output="command timed out" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.788243 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.786698 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="d24c721d-4968-477f-ba8e-23f6d31629d2" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.801773 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.804703 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"bb348c2b124362aa1be574374c3cd321b25ea76e59aee8aba55f432d9c45538f"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Mar 17 17:01:24 crc kubenswrapper[4767]: I0317 17:01:24.804843 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d24c721d-4968-477f-ba8e-23f6d31629d2" containerName="ceilometer-central-agent" containerID="cri-o://bb348c2b124362aa1be574374c3cd321b25ea76e59aee8aba55f432d9c45538f" gracePeriod=30 Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.118886 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" podUID="a7b5a1f5-da92-46dd-a5b0-5088e75346af" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.420598 4767 patch_prober.go:28] interesting pod/logging-loki-query-frontend-ff66c4dc9-4jtkt container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.420636 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-767865f676-btzsh" podUID="8bfc70f3-ead2-43cc-9e90-7df32804d9ac" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.420702 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" podUID="ce1ebb9b-1558-4190-9e47-4ef6ba3160da" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.435706 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerDied","Data":"fbfb64eda9a1857b0e600f098e323af2c36d82414c8ce1e3bffb3520a89ba819"} Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.436819 4767 generic.go:334] "Generic (PLEG): container finished" podID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerID="fbfb64eda9a1857b0e600f098e323af2c36d82414c8ce1e3bffb3520a89ba819" exitCode=143 Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.502500 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xfv84" podUID="81201888-f7c3-4605-86a7-85f3edbca3a6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.502560 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" podUID="43b467d2-9860-4feb-a656-65827836c23e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.502585 4767 patch_prober.go:28] interesting pod/logging-loki-distributor-9c6b6d984-q9dnr container/loki-distributor namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.502682 4767 patch_prober.go:28] interesting pod/metrics-server-56684ddd8d-zrdxq container/metrics-server namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.84:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.502691 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" podUID="28ac74bb-3f68-478d-8f50-acf61b2e8223" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.53:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.502610 4767 patch_prober.go:28] interesting pod/metrics-server-56684ddd8d-zrdxq container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.84:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.502721 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" podUID="5a2cfb18-0930-434f-a9bc-824575209dda" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.84:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.502824 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-56684ddd8d-zrdxq" podUID="5a2cfb18-0930-434f-a9bc-824575209dda" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.84:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.594528 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" podUID="921c5e77-9858-4177-99a0-8cd3a4420d7d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.717483 4767 patch_prober.go:28] interesting pod/observability-operator-6dd7dd855f-hw4fk container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.717567 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" podUID="bde9e722-7f00-444e-92fe-ac0db62b6172" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.717558 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-884679f54-44q8c" podUID="a833dc16-3a29-4129-b592-732d71818bc4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.717595 4767 patch_prober.go:28] interesting pod/observability-operator-6dd7dd855f-hw4fk container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.717699 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-6dd7dd855f-hw4fk" podUID="bde9e722-7f00-444e-92fe-ac0db62b6172" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.717529 4767 patch_prober.go:28] interesting pod/logging-loki-querier-6dcbdf8bb8-679nd container/loki-querier namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.54:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.717747 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" podUID="66451b46-d0f5-4037-aa4d-abf0143b2ef8" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.54:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.753282 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output="command timed out" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.755901 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output="command timed out" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.759799 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-5784578c99-2k2dh" podUID="bd2dbd72-69bf-40c1-b591-be3782c33465" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.759884 4767 patch_prober.go:28] interesting pod/monitoring-plugin-6bbd5f6758-kwcw5 container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.85:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.759913 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" podUID="93033ac4-fd81-46eb-8014-184056ab6de2" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.85:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.849427 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" podUID="59ba3b25-1e95-41ad-921c-9ee4ec5e2c43" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.977395 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-766dd46678-n9j9q" podUID="7381ff68-5e5d-4281-9924-8495ece760f4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:25 crc kubenswrapper[4767]: I0317 17:01:25.977395 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" podUID="e59ff072-51c2-4995-9f61-709f12a1393a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:25.977555 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" podUID="ad853953-ec48-40fc-8787-b2b838c955e9" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.082359 4767 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.082800 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="89ce8746-1721-4dec-9742-0abc2f498de3" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.123618 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.123669 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:8081/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.123706 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.123745 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.56:8081/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.137806 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-8q5h4 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.137891 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podUID="560edfa4-8a88-4c9b-8b31-e61f93050c15" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.137929 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-8q5h4 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.138015 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podUID="560edfa4-8a88-4c9b-8b31-e61f93050c15" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.157067 4767 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.157127 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a2389106-04cc-4871-b544-f8a677a6405a" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.58:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.166008 4767 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.166090 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="27c143b3-0f5b-4265-a2e2-d4302a2f3c70" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.288410 4767 patch_prober.go:28] interesting pod/perses-operator-6d5d56fd99-rxwwn container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.23:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.288522 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-6d5d56fd99-rxwwn" podUID="2531b56b-9f38-4cef-9812-b1b30c29405a" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.23:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.386382 4767 patch_prober.go:28] interesting pod/logging-loki-query-frontend-ff66c4dc9-4jtkt container/loki-query-frontend namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.386485 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-4jtkt" podUID="ce1ebb9b-1558-4190-9e47-4ef6ba3160da" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.483745 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" podUID="faf77dfd-c76e-4abd-a8c8-b6211cf070a4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.125:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.483920 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.620717 4767 patch_prober.go:28] interesting pod/thanos-querier-89d788fd9-2skkr container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.82:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:26 crc kubenswrapper[4767]: I0317 17:01:26.620835 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-89d788fd9-2skkr" podUID="8ab6a6b4-cb73-4701-a29c-f404c58966c4" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.82:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.056693 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.061416 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.057222 4767 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-pv6ds container/oauth-apiserver namespace/openshift-oauth-apiserver: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.061522 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pv6ds" podUID="ece0e760-026e-4de3-bb90-bb5117963f69" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.057275 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.061891 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.081788 4767 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.61:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.081865 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="89ce8746-1721-4dec-9742-0abc2f498de3" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.61:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.155136 4767 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.58:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.155254 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-ingester-0" podUID="a2389106-04cc-4871-b544-f8a677a6405a" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.58:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.157437 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-wjz85" podUID="1cfcd066-3d8a-431e-a895-a257a3e1baef" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.157610 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-wjz85" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.167652 4767 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.60:3101/loki/api/v1/status/buildinfo\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.167784 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-compactor-0" podUID="27c143b3-0f5b-4265-a2e2-d4302a2f3c70" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.60:3101/loki/api/v1/status/buildinfo\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.198511 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-wjz85" podUID="1cfcd066-3d8a-431e-a895-a257a3e1baef" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.198577 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/gateway namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.56:8081/live\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.198620 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/speaker-wjz85" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.198653 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.56:8081/live\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.198674 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/opa namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.56:8083/live\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.198744 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-8q5h4 container/opa namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.57:8083/live\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.198765 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.56:8083/live\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.198807 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-8q5h4 container/gateway namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.57:8081/live\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.198781 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podUID="560edfa4-8a88-4c9b-8b31-e61f93050c15" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/live\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.198829 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podUID="560edfa4-8a88-4c9b-8b31-e61f93050c15" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/live\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.377967 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="94b1e014-c27b-4b81-a12d-eecfe02d5b67" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.378768 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="94b1e014-c27b-4b81-a12d-eecfe02d5b67" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.472276 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerStarted","Data":"799ebbbcff35e11ae692c17c0fc648dd6524899d0215b833b7a14d1095ff8b0b"} Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.496533 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="speaker" containerStatusID={"Type":"cri-o","ID":"e809b31439dc0f3d987c0a5d250a1921743dd5246b80134e8aa65b80a195605d"} pod="metallb-system/speaker-wjz85" containerMessage="Container speaker failed liveness probe, will be restarted" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.498067 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/speaker-wjz85" podUID="1cfcd066-3d8a-431e-a895-a257a3e1baef" containerName="speaker" containerID="cri-o://e809b31439dc0f3d987c0a5d250a1921743dd5246b80134e8aa65b80a195605d" gracePeriod=2 Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.539468 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" podUID="faf77dfd-c76e-4abd-a8c8-b6211cf070a4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.125:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.895692 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.895692 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.895796 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.895926 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.896008 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.896046 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.897730 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="console-operator" containerStatusID={"Type":"cri-o","ID":"0adbf15d1072c1fe23beb1443ff1648323f80fd1ab7345b545ce8b2cdd2060e7"} pod="openshift-console-operator/console-operator-58897d9998-rbpmk" containerMessage="Container console-operator failed liveness probe, will be restarted" Mar 17 17:01:27 crc kubenswrapper[4767]: I0317 17:01:27.897813 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" containerID="cri-o://0adbf15d1072c1fe23beb1443ff1648323f80fd1ab7345b545ce8b2cdd2060e7" gracePeriod=30 Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.055028 4767 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-mdlms container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.055114 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" podUID="a8829d5d-100e-4518-8863-db3ab9c7b30d" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.055202 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.057008 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="authentication-operator" containerStatusID={"Type":"cri-o","ID":"3942c07c07c5367fe53baeedc2dbb54cea88cdea4ebf87f7f586fc0bf82810ae"} pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" containerMessage="Container authentication-operator failed liveness probe, will be restarted" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.057079 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" podUID="a8829d5d-100e-4518-8863-db3ab9c7b30d" containerName="authentication-operator" containerID="cri-o://3942c07c07c5367fe53baeedc2dbb54cea88cdea4ebf87f7f586fc0bf82810ae" gracePeriod=30 Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.200688 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-wjz85" podUID="1cfcd066-3d8a-431e-a895-a257a3e1baef" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.751960 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="60ee8af2-4ae0-402e-8a13-2dce36db4bab" containerName="prometheus" probeResult="failure" output="command timed out" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.755014 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-mdghs" podUID="571d124c-f5cb-4c5e-bfd7-3f56836ce800" containerName="registry-server" probeResult="failure" output="command timed out" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.755497 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-mdghs" podUID="571d124c-f5cb-4c5e-bfd7-3f56836ce800" containerName="registry-server" probeResult="failure" output="command timed out" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.831286 4767 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-82k6b container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.831356 4767 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-82k6b container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.831378 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.831484 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.831484 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.831568 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.833287 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="prometheus-operator-admission-webhook" containerStatusID={"Type":"cri-o","ID":"21bd32970bf467e403774e6b15ad900b53bef84f68891cc87f2251b74697e862"} pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" containerMessage="Container prometheus-operator-admission-webhook failed liveness probe, will be restarted" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.833330 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" containerID="cri-o://21bd32970bf467e403774e6b15ad900b53bef84f68891cc87f2251b74697e862" gracePeriod=30 Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.897086 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.897229 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:28 crc kubenswrapper[4767]: I0317 17:01:28.994710 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-sl4m8" podUID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:28 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:28 crc kubenswrapper[4767]: > Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.083289 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-hqqbp" podUID="349fc544-d564-4b99-bac0-bfba49fe72db" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:29 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:29 crc kubenswrapper[4767]: > Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.242375 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-wjz85" podUID="1cfcd066-3d8a-431e-a895-a257a3e1baef" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.354672 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.355041 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.355560 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-59hsl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.355740 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-59hsl" podUID="471b6054-3b2a-47c9-8889-942603600da9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.539007 4767 patch_prober.go:28] interesting pod/apiserver-76f77b778f-bk7p5 container/openshift-apiserver namespace/openshift-apiserver: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/readyz?exclude=etcd&exclude=etcd-readiness\": context deadline exceeded" start-of-body= Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.539088 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" podUID="1c88a6ee-cd24-4d85-8a89-d830e5baa434" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.36:8443/readyz?exclude=etcd&exclude=etcd-readiness\": context deadline exceeded" Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.539978 4767 patch_prober.go:28] interesting pod/apiserver-76f77b778f-bk7p5 container/openshift-apiserver namespace/openshift-apiserver: Liveness probe status=failure output="Get \"https://10.217.0.36:8443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.540060 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-apiserver/apiserver-76f77b778f-bk7p5" podUID="1c88a6ee-cd24-4d85-8a89-d830e5baa434" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.36:8443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.750604 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-sl4m8" podUID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerName="registry-server" probeResult="failure" output="command timed out" Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.752116 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="60ee8af2-4ae0-402e-8a13-2dce36db4bab" containerName="prometheus" probeResult="failure" output="command timed out" Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.787644 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-5tr6k" podUID="cc1f08cb-7205-4537-acca-91edee17aef9" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:29 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:29 crc kubenswrapper[4767]: > Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.787674 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-hqqbp" podUID="349fc544-d564-4b99-bac0-bfba49fe72db" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:29 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:29 crc kubenswrapper[4767]: > Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.787842 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-5tr6k" podUID="cc1f08cb-7205-4537-acca-91edee17aef9" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:29 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:29 crc kubenswrapper[4767]: > Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.832842 4767 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-82k6b container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.832941 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.76:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:29 crc kubenswrapper[4767]: E0317 17:01:29.908664 4767 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.948500 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" podUID="95d5d3e8-dc72-414f-afe4-b68b757a39d4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.948588 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" podUID="95d5d3e8-dc72-414f-afe4-b68b757a39d4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:29 crc kubenswrapper[4767]: I0317 17:01:29.948947 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.099206 4767 trace.go:236] Trace[581096676]: "Calculate volume metrics of storage for pod openshift-logging/logging-loki-compactor-0" (17-Mar-2026 17:01:27.463) (total time: 2628ms): Mar 17 17:01:30 crc kubenswrapper[4767]: Trace[581096676]: [2.628838792s] [2.628838792s] END Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.138412 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.139591 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.138443 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.139817 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.139936 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.139969 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.141522 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="openshift-config-operator" containerStatusID={"Type":"cri-o","ID":"d1e085d042d9d8b5cec9ab40cfe208b9224e0fe81e8ca191bf191605cd7af4c7"} pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" containerMessage="Container openshift-config-operator failed liveness probe, will be restarted" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.141573 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" containerID="cri-o://d1e085d042d9d8b5cec9ab40cfe208b9224e0fe81e8ca191bf191605cd7af4c7" gracePeriod=30 Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.280964 4767 trace.go:236] Trace[2137130999]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-2" (17-Mar-2026 17:01:26.762) (total time: 3518ms): Mar 17 17:01:30 crc kubenswrapper[4767]: Trace[2137130999]: [3.518878781s] [3.518878781s] END Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.348517 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.348631 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.348781 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.348514 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.349292 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.349499 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.351282 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="router" containerStatusID={"Type":"cri-o","ID":"3dbc65c86754010059515f8fc9873782ca67f8577939e37319a0dcd29bf71113"} pod="openshift-ingress/router-default-5444994796-ztnfb" containerMessage="Container router failed liveness probe, will be restarted" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.351366 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" containerID="cri-o://3dbc65c86754010059515f8fc9873782ca67f8577939e37319a0dcd29bf71113" gracePeriod=10 Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.372762 4767 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:6443/livez?exclude=etcd\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.372858 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez?exclude=etcd\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.373251 4767 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.373326 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.541824 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wjz85" event={"ID":"1cfcd066-3d8a-431e-a895-a257a3e1baef","Type":"ContainerDied","Data":"e809b31439dc0f3d987c0a5d250a1921743dd5246b80134e8aa65b80a195605d"} Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.545184 4767 generic.go:334] "Generic (PLEG): container finished" podID="1cfcd066-3d8a-431e-a895-a257a3e1baef" containerID="e809b31439dc0f3d987c0a5d250a1921743dd5246b80134e8aa65b80a195605d" exitCode=137 Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.580404 4767 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-lzjm6 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.580460 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.580503 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" podUID="68f6ff1d-a119-4faf-89aa-fd9afc2d93f8" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.580564 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.580539 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.580617 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.580783 4767 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-lzjm6 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.580806 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" podUID="68f6ff1d-a119-4faf-89aa-fd9afc2d93f8" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.580852 4767 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.580870 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581395 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581418 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581454 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581466 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.580425 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581507 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581773 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581814 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581831 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581839 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581849 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581911 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="package-server-manager" containerStatusID={"Type":"cri-o","ID":"4965483e6277ff80bc871b75394d2f20f37e0b634fd01578030291d8d29bfe8a"} pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" containerMessage="Container package-server-manager failed liveness probe, will be restarted" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.581971 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" podUID="68f6ff1d-a119-4faf-89aa-fd9afc2d93f8" containerName="package-server-manager" containerID="cri-o://4965483e6277ff80bc871b75394d2f20f37e0b634fd01578030291d8d29bfe8a" gracePeriod=30 Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.597461 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="olm-operator" containerStatusID={"Type":"cri-o","ID":"52a6acfe67114f9acc7cfa43c05c4e8be607930c51ff296233e402842a460fb6"} pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" containerMessage="Container olm-operator failed liveness probe, will be restarted" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.597566 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" containerID="cri-o://52a6acfe67114f9acc7cfa43c05c4e8be607930c51ff296233e402842a460fb6" gracePeriod=30 Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.601131 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="catalog-operator" containerStatusID={"Type":"cri-o","ID":"59c17ad5134f8551f9dd65e38fcb77e2bca464ade2812471744070029473321e"} pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" containerMessage="Container catalog-operator failed liveness probe, will be restarted" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.601257 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" containerID="cri-o://59c17ad5134f8551f9dd65e38fcb77e2bca464ade2812471744070029473321e" gracePeriod=30 Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.605642 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-scheduler" containerStatusID={"Type":"cri-o","ID":"ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb"} pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" containerMessage="Container kube-scheduler failed liveness probe, will be restarted" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.605759 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" containerID="cri-o://ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb" gracePeriod=30 Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.831562 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-rbpmk_3e0ef009-3626-492a-828e-be5f7285227c/console-operator/0.log" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.831756 4767 generic.go:334] "Generic (PLEG): container finished" podID="3e0ef009-3626-492a-828e-be5f7285227c" containerID="0adbf15d1072c1fe23beb1443ff1648323f80fd1ab7345b545ce8b2cdd2060e7" exitCode=1 Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.831849 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" event={"ID":"3e0ef009-3626-492a-828e-be5f7285227c","Type":"ContainerDied","Data":"0adbf15d1072c1fe23beb1443ff1648323f80fd1ab7345b545ce8b2cdd2060e7"} Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.990558 4767 patch_prober.go:28] interesting pod/oauth-openshift-5cc5b65bd-6nll7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.990593 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" podUID="95d5d3e8-dc72-414f-afe4-b68b757a39d4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.991030 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" podUID="2321f9fb-78c8-401b-8fc2-a5f82d42a226" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.990679 4767 patch_prober.go:28] interesting pod/oauth-openshift-5cc5b65bd-6nll7 container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.991116 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" podUID="2321f9fb-78c8-401b-8fc2-a5f82d42a226" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.991165 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.990732 4767 patch_prober.go:28] interesting pod/controller-manager-57bc79d78-25d96 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.991219 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.991234 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podUID="1767bf72-c95d-486d-baa6-ca741742569b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.991293 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.990760 4767 patch_prober.go:28] interesting pod/controller-manager-57bc79d78-25d96 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.991404 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podUID="1767bf72-c95d-486d-baa6-ca741742569b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.993575 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="oauth-openshift" containerStatusID={"Type":"cri-o","ID":"eaff0e00c90c96aca126c5718dbe75b4480b9cf9712683c7700af0e35030c21d"} pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" containerMessage="Container oauth-openshift failed liveness probe, will be restarted" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.993676 4767 patch_prober.go:28] interesting pod/route-controller-manager-56d7bf8c9b-cs8b8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.993841 4767 patch_prober.go:28] interesting pod/route-controller-manager-56d7bf8c9b-cs8b8 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.993886 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" podUID="a9f5b566-a215-4322-8d74-668dc2a93f3e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:30 crc kubenswrapper[4767]: I0317 17:01:30.993940 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:30.993706 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" podUID="a9f5b566-a215-4322-8d74-668dc2a93f3e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.003489 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="route-controller-manager" containerStatusID={"Type":"cri-o","ID":"7568ac5efdde80256b6a21b3c20a6b083df449905aa94a3be7af802805f2636f"} pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" containerMessage="Container route-controller-manager failed liveness probe, will be restarted" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.003516 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="controller-manager" containerStatusID={"Type":"cri-o","ID":"07a6477a1b291c54a3879e589f04d6266ea243a0214e6d194b7b8b54cd46e3f2"} pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" containerMessage="Container controller-manager failed liveness probe, will be restarted" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.003644 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podUID="1767bf72-c95d-486d-baa6-ca741742569b" containerName="controller-manager" containerID="cri-o://07a6477a1b291c54a3879e589f04d6266ea243a0214e6d194b7b8b54cd46e3f2" gracePeriod=30 Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.003827 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" podUID="a9f5b566-a215-4322-8d74-668dc2a93f3e" containerName="route-controller-manager" containerID="cri-o://7568ac5efdde80256b6a21b3c20a6b083df449905aa94a3be7af802805f2636f" gracePeriod=30 Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.117507 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.117595 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.117614 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.117660 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.117687 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.117731 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.119236 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="packageserver" containerStatusID={"Type":"cri-o","ID":"17d8403e3f262fd23c063dcbbb20f9f6124c19c1851c67e125a6ab143e33581f"} pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" containerMessage="Container packageserver failed liveness probe, will be restarted" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.119289 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" containerID="cri-o://17d8403e3f262fd23c063dcbbb20f9f6124c19c1851c67e125a6ab143e33581f" gracePeriod=30 Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.124265 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.124319 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-7gw68 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.124354 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.56:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.124316 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-7gw68" podUID="9c622cae-f125-4675-9bb5-d2dd999edfa3" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.137235 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-8q5h4 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.137317 4767 patch_prober.go:28] interesting pod/logging-loki-gateway-76dddc4d57-8q5h4 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.137419 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podUID="560edfa4-8a88-4c9b-8b31-e61f93050c15" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.137310 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-76dddc4d57-8q5h4" podUID="560edfa4-8a88-4c9b-8b31-e61f93050c15" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.389581 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.389672 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.430529 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.430595 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.536359 4767 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.536480 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" podUID="70fbe8aa-8647-4fe8-914c-0c05399cf46d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.536980 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.536566 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" podUID="70fbe8aa-8647-4fe8-914c-0c05399cf46d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.537105 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.541039 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.584147 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.584241 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.749800 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="60ee8af2-4ae0-402e-8a13-2dce36db4bab" containerName="prometheus" probeResult="failure" output="command timed out" Mar 17 17:01:31 crc kubenswrapper[4767]: I0317 17:01:31.992096 4767 patch_prober.go:28] interesting pod/oauth-openshift-5cc5b65bd-6nll7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:31.992193 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" podUID="2321f9fb-78c8-401b-8fc2-a5f82d42a226" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.056437 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.056528 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.151368 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" podUID="713154c5-7e16-498f-9612-1c0afbf362bb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.151389 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" podUID="713154c5-7e16-498f-9612-1c0afbf362bb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.151433 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.151868 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.151984 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.196430 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" podUID="877f4eda-1ec7-4296-98df-b5ca7a7fa78a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.96:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.196577 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.377314 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="94b1e014-c27b-4b81-a12d-eecfe02d5b67" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.377366 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="94b1e014-c27b-4b81-a12d-eecfe02d5b67" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.377518 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.419833 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-wj8xp" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.563447 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" podUID="fda1c393-424a-4142-8570-5dc108f6d6f4" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.563591 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.606472 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" podUID="fda1c393-424a-4142-8570-5dc108f6d6f4" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.606585 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.606640 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" podUID="70fbe8aa-8647-4fe8-914c-0c05399cf46d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.695095 4767 patch_prober.go:28] interesting pod/etcd-crc container/etcd namespace/openshift-etcd: Liveness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=failed to establish etcd client: giving up getting a cached client after 3 tries Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.695208 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-etcd/etcd-crc" podUID="2139d3e2895fc6797b9c76a1b4c9886d" containerName="etcd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.695747 4767 patch_prober.go:28] interesting pod/etcd-crc container/etcd namespace/openshift-etcd: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=failed to establish etcd client: giving up getting a cached client after 3 tries Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.695798 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-etcd/etcd-crc" podUID="2139d3e2895fc6797b9c76a1b4c9886d" containerName="etcd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.826438 4767 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.826529 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.873973 4767 generic.go:334] "Generic (PLEG): container finished" podID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerID="59c17ad5134f8551f9dd65e38fcb77e2bca464ade2812471744070029473321e" exitCode=0 Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.874048 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" event={"ID":"75f035cd-32d9-44a2-8d8f-3d5f40742b67","Type":"ContainerDied","Data":"59c17ad5134f8551f9dd65e38fcb77e2bca464ade2812471744070029473321e"} Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.874558 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="webhook-server" containerStatusID={"Type":"cri-o","ID":"f8be811582075c48d993dad19d81c0727dfffbb03d827ac409ec6fe280626d48"} pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" containerMessage="Container webhook-server failed liveness probe, will be restarted" Mar 17 17:01:32 crc kubenswrapper[4767]: I0317 17:01:32.874600 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" podUID="fda1c393-424a-4142-8570-5dc108f6d6f4" containerName="webhook-server" containerID="cri-o://f8be811582075c48d993dad19d81c0727dfffbb03d827ac409ec6fe280626d48" gracePeriod=2 Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.105927 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-wj8xp" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.195076 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" podUID="713154c5-7e16-498f-9612-1c0afbf362bb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.239506 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" podUID="877f4eda-1ec7-4296-98df-b5ca7a7fa78a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.96:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.291729 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.394888 4767 patch_prober.go:28] interesting pod/console-89866dfb6-fswnc container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.146:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.395016 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-89866dfb6-fswnc" podUID="b52b7ad3-98cb-4051-8e3b-665fa44b0fd6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.146:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.395141 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-89866dfb6-fswnc" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.502466 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-wj8xp" podUID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.502913 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-wj8xp" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.585525 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" podUID="07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.585653 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.585954 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" podUID="07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.586103 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.586105 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-wj8xp" podUID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.586200 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/frr-k8s-wj8xp" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.586912 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="frr-k8s-webhook-server" containerStatusID={"Type":"cri-o","ID":"419cc940f05664f6f7e89224b92ed5c9bd7ed965f7d6d4de25d564d24a720843"} pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" containerMessage="Container frr-k8s-webhook-server failed liveness probe, will be restarted" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.586966 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" podUID="07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec" containerName="frr-k8s-webhook-server" containerID="cri-o://419cc940f05664f6f7e89224b92ed5c9bd7ed965f7d6d4de25d564d24a720843" gracePeriod=10 Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.626715 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" podUID="fda1c393-424a-4142-8570-5dc108f6d6f4" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.677477 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-init-587c5db989-zt4g2" podUID="3ee9406d-d284-417c-83e2-942103e449eb" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.750294 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.768485 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-kv9bb" podUID="baf4c43e-08e2-4afa-b9b5-8bb4aa8f0352" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.901628 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wjz85" event={"ID":"1cfcd066-3d8a-431e-a895-a257a3e1baef","Type":"ContainerStarted","Data":"31eeb822267b5f63a9b4aabf791526f889119ad65bbf55efe45b7f272840b650"} Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.901974 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-wjz85" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.906065 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.912822 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.918102 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.918207 4767 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="1a0bfdf7badf045b41c62ef25a6c37928c0abf036ce202175f2dd3cbc88a647c" exitCode=1 Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.918298 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"1a0bfdf7badf045b41c62ef25a6c37928c0abf036ce202175f2dd3cbc88a647c"} Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.919078 4767 scope.go:117] "RemoveContainer" containerID="1a6d14de833771c3a060a2879802f4f7f37b96b60004b4e5f96374e5fee1428e" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.920101 4767 scope.go:117] "RemoveContainer" containerID="1a0bfdf7badf045b41c62ef25a6c37928c0abf036ce202175f2dd3cbc88a647c" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.922019 4767 generic.go:334] "Generic (PLEG): container finished" podID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerID="21bd32970bf467e403774e6b15ad900b53bef84f68891cc87f2251b74697e862" exitCode=0 Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.922100 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" event={"ID":"15e74b29-0b89-4e6a-84aa-94d9f5faa0e5","Type":"ContainerDied","Data":"21bd32970bf467e403774e6b15ad900b53bef84f68891cc87f2251b74697e862"} Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.929415 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-rbpmk_3e0ef009-3626-492a-828e-be5f7285227c/console-operator/0.log" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.929549 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" event={"ID":"3e0ef009-3626-492a-828e-be5f7285227c","Type":"ContainerStarted","Data":"92884695dfb1c97e0192d7121021a30acdc67e8fbfbb6da11ae403afe2dfd976"} Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.930447 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="controller" containerStatusID={"Type":"cri-o","ID":"b3c9ce2a1c6b819b46adecf1b0af9f33b68e25920278ace693bc6bd14479bdf5"} pod="metallb-system/frr-k8s-wj8xp" containerMessage="Container controller failed liveness probe, will be restarted" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.930569 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-wj8xp" podUID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerName="controller" containerID="cri-o://b3c9ce2a1c6b819b46adecf1b0af9f33b68e25920278ace693bc6bd14479bdf5" gracePeriod=2 Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.930759 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.933507 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 17 17:01:33 crc kubenswrapper[4767]: I0317 17:01:33.933558 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.040497 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" podUID="53aea289-5a28-438b-8d28-242d836351f0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.040647 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.040871 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" podUID="53aea289-5a28-438b-8d28-242d836351f0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.237460 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-7bb4cc7c98-nchz8" podUID="b978bf52-fdb5-4863-a5bf-e2aec8ab08a4" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.99:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.237867 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-wj8xp" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.237552 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-7bb4cc7c98-nchz8" podUID="b978bf52-fdb5-4863-a5bf-e2aec8ab08a4" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.99:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.237876 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.237690 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" podUID="fda1c393-424a-4142-8570-5dc108f6d6f4" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": EOF" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.237984 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.284222 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.346504 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-q7zd5" podUID="a67a8ddc-0801-4530-bf82-9d4f4d0389a8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.396822 4767 patch_prober.go:28] interesting pod/console-89866dfb6-fswnc container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.146:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.396948 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-89866dfb6-fswnc" podUID="b52b7ad3-98cb-4051-8e3b-665fa44b0fd6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.146:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.442400 4767 patch_prober.go:28] interesting pod/nmstate-webhook-5f558f5558-rwq8w container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.89:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.442473 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" podUID="31f30608-3b0d-4f63-9ab8-6a1547e233ff" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.89:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.442561 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.485471 4767 patch_prober.go:28] interesting pod/loki-operator-controller-manager-5dcc96cb6-rv7xm container/manager namespace/openshift-operators-redhat: Liveness probe status=failure output="Get \"http://10.217.0.50:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.485835 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" podUID="6f682203-ee25-4b31-957b-26d8148feda0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.50:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.567439 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" podUID="0163654c-d57e-4b14-aba0-f76dbaff1114" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.567507 4767 patch_prober.go:28] interesting pod/logging-loki-distributor-9c6b6d984-q9dnr container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.567579 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" podUID="28ac74bb-3f68-478d-8f50-acf61b2e8223" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.53:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.567715 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.567947 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-drx6n" podUID="0163654c-d57e-4b14-aba0-f76dbaff1114" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.568560 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-rwq8w" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.588672 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" podUID="904d9f3b-95f0-4e57-8d04-c2026227a4a6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.657801 4767 patch_prober.go:28] interesting pod/logging-loki-querier-6dcbdf8bb8-679nd container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.657898 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" podUID="66451b46-d0f5-4037-aa4d-abf0143b2ef8" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.658226 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.750667 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="6abe028c-416e-4978-bcc8-3a7b8d92624b" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.750827 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.753282 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"6abac3a757369ed7aa38da0ae38fc79753f588411284852e6d063714ed21a5fd"} pod="openstack/openstack-cell1-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Mar 17 17:01:34 crc kubenswrapper[4767]: E0317 17:01:34.794634 4767 controller.go:195] "Failed to update lease" err="Operation cannot be fulfilled on leases.coordination.k8s.io \"crc\": the object has been modified; please apply your changes to the latest version and try again" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.800401 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-rvkxc" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.850538 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-9c6b6d984-q9dnr" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.878199 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-679nd" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.946031 4767 generic.go:334] "Generic (PLEG): container finished" podID="904d9f3b-95f0-4e57-8d04-c2026227a4a6" containerID="1e47fd04dd013933d9e3c724fbf14b32d6b4c2a21b0d2c1b23dd76b6abb46f2f" exitCode=1 Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.946116 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" event={"ID":"904d9f3b-95f0-4e57-8d04-c2026227a4a6","Type":"ContainerDied","Data":"1e47fd04dd013933d9e3c724fbf14b32d6b4c2a21b0d2c1b23dd76b6abb46f2f"} Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.950667 4767 scope.go:117] "RemoveContainer" containerID="1e47fd04dd013933d9e3c724fbf14b32d6b4c2a21b0d2c1b23dd76b6abb46f2f" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.953863 4767 generic.go:334] "Generic (PLEG): container finished" podID="e5da4ccd-4ecb-4929-8974-0a31d018204d" containerID="b3c9ce2a1c6b819b46adecf1b0af9f33b68e25920278ace693bc6bd14479bdf5" exitCode=0 Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.953940 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerDied","Data":"b3c9ce2a1c6b819b46adecf1b0af9f33b68e25920278ace693bc6bd14479bdf5"} Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.957677 4767 generic.go:334] "Generic (PLEG): container finished" podID="a333e040-ed72-4d18-95b1-14b57a92c895" containerID="52a6acfe67114f9acc7cfa43c05c4e8be607930c51ff296233e402842a460fb6" exitCode=0 Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.958295 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" event={"ID":"a333e040-ed72-4d18-95b1-14b57a92c895","Type":"ContainerDied","Data":"52a6acfe67114f9acc7cfa43c05c4e8be607930c51ff296233e402842a460fb6"} Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.970588 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" event={"ID":"15e74b29-0b89-4e6a-84aa-94d9f5faa0e5","Type":"ContainerStarted","Data":"f23677de1f21f25ccec2a570df9f273571bff422250784c23831123d7d2693fe"} Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.970810 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.971299 4767 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-82k6b container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.76:8443/healthz\": dial tcp 10.217.0.76:8443: connect: connection refused" start-of-body= Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.971353 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.76:8443/healthz\": dial tcp 10.217.0.76:8443: connect: connection refused" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.973675 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" event={"ID":"75f035cd-32d9-44a2-8d8f-3d5f40742b67","Type":"ContainerStarted","Data":"39b907809971f63ba11a5cc890121bea80236fbdff3326187d9ed964b55ef530"} Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.974907 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.975474 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.975528 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.980611 4767 generic.go:334] "Generic (PLEG): container finished" podID="fda1c393-424a-4142-8570-5dc108f6d6f4" containerID="f8be811582075c48d993dad19d81c0727dfffbb03d827ac409ec6fe280626d48" exitCode=0 Mar 17 17:01:34 crc kubenswrapper[4767]: I0317 17:01:34.980755 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" event={"ID":"fda1c393-424a-4142-8570-5dc108f6d6f4","Type":"ContainerDied","Data":"f8be811582075c48d993dad19d81c0727dfffbb03d827ac409ec6fe280626d48"} Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.013390 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.015631 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.025946 4767 generic.go:334] "Generic (PLEG): container finished" podID="70fbe8aa-8647-4fe8-914c-0c05399cf46d" containerID="f42bd3c4f2af2dd50de480c34d99956b8b583c9a3551b160871025ce4d7b5666" exitCode=1 Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.026055 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" event={"ID":"70fbe8aa-8647-4fe8-914c-0c05399cf46d","Type":"ContainerDied","Data":"f42bd3c4f2af2dd50de480c34d99956b8b583c9a3551b160871025ce4d7b5666"} Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.027312 4767 scope.go:117] "RemoveContainer" containerID="f42bd3c4f2af2dd50de480c34d99956b8b583c9a3551b160871025ce4d7b5666" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.029870 4767 generic.go:334] "Generic (PLEG): container finished" podID="07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec" containerID="419cc940f05664f6f7e89224b92ed5c9bd7ed965f7d6d4de25d564d24a720843" exitCode=0 Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.030455 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" event={"ID":"07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec","Type":"ContainerDied","Data":"419cc940f05664f6f7e89224b92ed5c9bd7ed965f7d6d4de25d564d24a720843"} Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.032040 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.032090 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.056151 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.056258 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.386323 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-sl4m8" podUID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:35 crc kubenswrapper[4767]: timeout: health rpc did not complete within 1s Mar 17 17:01:35 crc kubenswrapper[4767]: > Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.395811 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:35 crc kubenswrapper[4767]: timeout: health rpc did not complete within 1s Mar 17 17:01:35 crc kubenswrapper[4767]: > Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.405513 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:35 crc kubenswrapper[4767]: timeout: health rpc did not complete within 1s Mar 17 17:01:35 crc kubenswrapper[4767]: > Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.448829 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" podUID="43b467d2-9860-4feb-a656-65827836c23e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.497024 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" podUID="43b467d2-9860-4feb-a656-65827836c23e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.517467 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-sl4m8" podUID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:35 crc kubenswrapper[4767]: timeout: health rpc did not complete within 1s Mar 17 17:01:35 crc kubenswrapper[4767]: > Mar 17 17:01:35 crc kubenswrapper[4767]: E0317 17:01:35.728207 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T17:01:25Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T17:01:25Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T17:01:25Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T17:01:25Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.735371 4767 patch_prober.go:28] interesting pod/monitoring-plugin-6bbd5f6758-kwcw5 container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.85:9443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.735441 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" podUID="93033ac4-fd81-46eb-8014-184056ab6de2" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.85:9443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.788793 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="6abe028c-416e-4978-bcc8-3a7b8d92624b" containerName="galera" probeResult="failure" output="command timed out" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.942519 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" podUID="ad853953-ec48-40fc-8787-b2b838c955e9" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.942634 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-c674c5965-jx2k5" podUID="ad853953-ec48-40fc-8787-b2b838c955e9" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.952582 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.952748 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.952790 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.952820 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.952859 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-22vp8" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.952962 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.952979 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.953058 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.960988 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"80cdce19b94103bc1982d3760e2748cdc19f5e1e69691384f05bd4ed1d8cf95e"} pod="openshift-marketplace/redhat-marketplace-sl4m8" containerMessage="Container registry-server failed liveness probe, will be restarted" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.962525 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sl4m8" podUID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerName="registry-server" containerID="cri-o://80cdce19b94103bc1982d3760e2748cdc19f5e1e69691384f05bd4ed1d8cf95e" gracePeriod=30 Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.963207 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"44e01702cdd6851f3e6eac45406e1af0302eab03207b5ef699665002ceb62b55"} pod="openshift-marketplace/redhat-operators-v5jdf" containerMessage="Container registry-server failed liveness probe, will be restarted" Mar 17 17:01:35 crc kubenswrapper[4767]: I0317 17:01:35.963280 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" containerID="cri-o://44e01702cdd6851f3e6eac45406e1af0302eab03207b5ef699665002ceb62b55" gracePeriod=30 Mar 17 17:01:35 crc kubenswrapper[4767]: E0317 17:01:35.980914 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="44e01702cdd6851f3e6eac45406e1af0302eab03207b5ef699665002ceb62b55" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.007785 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ks6jq" Mar 17 17:01:36 crc kubenswrapper[4767]: E0317 17:01:36.007940 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="44e01702cdd6851f3e6eac45406e1af0302eab03207b5ef699665002ceb62b55" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.013817 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-6bbd5f6758-kwcw5" Mar 17 17:01:36 crc kubenswrapper[4767]: E0317 17:01:36.030112 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="44e01702cdd6851f3e6eac45406e1af0302eab03207b5ef699665002ceb62b55" cmd=["grpc_health_probe","-addr=:50051"] Mar 17 17:01:36 crc kubenswrapper[4767]: E0317 17:01:36.030217 4767 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.117634 4767 generic.go:334] "Generic (PLEG): container finished" podID="6f682203-ee25-4b31-957b-26d8148feda0" containerID="e92b976b8865fc5db41829a9fa6f48341ac42c83d5319bca392cb1fe37e46e93" exitCode=1 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.117757 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" event={"ID":"6f682203-ee25-4b31-957b-26d8148feda0","Type":"ContainerDied","Data":"e92b976b8865fc5db41829a9fa6f48341ac42c83d5319bca392cb1fe37e46e93"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.119941 4767 scope.go:117] "RemoveContainer" containerID="e92b976b8865fc5db41829a9fa6f48341ac42c83d5319bca392cb1fe37e46e93" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.123251 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-sl4m8" podUID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerName="registry-server" probeResult="failure" output="" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.136439 4767 generic.go:334] "Generic (PLEG): container finished" podID="a9f5b566-a215-4322-8d74-668dc2a93f3e" containerID="7568ac5efdde80256b6a21b3c20a6b083df449905aa94a3be7af802805f2636f" exitCode=0 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.136533 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" event={"ID":"a9f5b566-a215-4322-8d74-668dc2a93f3e","Type":"ContainerDied","Data":"7568ac5efdde80256b6a21b3c20a6b083df449905aa94a3be7af802805f2636f"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.147303 4767 generic.go:334] "Generic (PLEG): container finished" podID="e59ff072-51c2-4995-9f61-709f12a1393a" containerID="40659e8a7dfbbdabdb7c5244845ba76a395f7824954f40eef8741026a2be3fe5" exitCode=1 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.147435 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" event={"ID":"e59ff072-51c2-4995-9f61-709f12a1393a","Type":"ContainerDied","Data":"40659e8a7dfbbdabdb7c5244845ba76a395f7824954f40eef8741026a2be3fe5"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.149797 4767 scope.go:117] "RemoveContainer" containerID="40659e8a7dfbbdabdb7c5244845ba76a395f7824954f40eef8741026a2be3fe5" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.151306 4767 generic.go:334] "Generic (PLEG): container finished" podID="a8829d5d-100e-4518-8863-db3ab9c7b30d" containerID="3942c07c07c5367fe53baeedc2dbb54cea88cdea4ebf87f7f586fc0bf82810ae" exitCode=0 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.151462 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" event={"ID":"a8829d5d-100e-4518-8863-db3ab9c7b30d","Type":"ContainerDied","Data":"3942c07c07c5367fe53baeedc2dbb54cea88cdea4ebf87f7f586fc0bf82810ae"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.154586 4767 generic.go:334] "Generic (PLEG): container finished" podID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerID="17d8403e3f262fd23c063dcbbb20f9f6124c19c1851c67e125a6ab143e33581f" exitCode=0 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.154673 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" event={"ID":"b836be84-e8b0-4e8d-8bac-cfc922bea7eb","Type":"ContainerDied","Data":"17d8403e3f262fd23c063dcbbb20f9f6124c19c1851c67e125a6ab143e33581f"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.157915 4767 generic.go:334] "Generic (PLEG): container finished" podID="59ba3b25-1e95-41ad-921c-9ee4ec5e2c43" containerID="4d4b716faf520d9a4ca657e0d7d163e03e65ffeb116ebaedfbe80d11d0bfe322" exitCode=1 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.157985 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" event={"ID":"59ba3b25-1e95-41ad-921c-9ee4ec5e2c43","Type":"ContainerDied","Data":"4d4b716faf520d9a4ca657e0d7d163e03e65ffeb116ebaedfbe80d11d0bfe322"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.158992 4767 scope.go:117] "RemoveContainer" containerID="4d4b716faf520d9a4ca657e0d7d163e03e65ffeb116ebaedfbe80d11d0bfe322" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.161569 4767 generic.go:334] "Generic (PLEG): container finished" podID="877f4eda-1ec7-4296-98df-b5ca7a7fa78a" containerID="b8a216894dbce2190c35e4629181d1975a45c2bddb4479f3c33d94755e92e1a2" exitCode=1 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.161637 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" event={"ID":"877f4eda-1ec7-4296-98df-b5ca7a7fa78a","Type":"ContainerDied","Data":"b8a216894dbce2190c35e4629181d1975a45c2bddb4479f3c33d94755e92e1a2"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.162477 4767 scope.go:117] "RemoveContainer" containerID="b8a216894dbce2190c35e4629181d1975a45c2bddb4479f3c33d94755e92e1a2" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.168649 4767 generic.go:334] "Generic (PLEG): container finished" podID="1767bf72-c95d-486d-baa6-ca741742569b" containerID="07a6477a1b291c54a3879e589f04d6266ea243a0214e6d194b7b8b54cd46e3f2" exitCode=0 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.168734 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" event={"ID":"1767bf72-c95d-486d-baa6-ca741742569b","Type":"ContainerDied","Data":"07a6477a1b291c54a3879e589f04d6266ea243a0214e6d194b7b8b54cd46e3f2"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.178002 4767 generic.go:334] "Generic (PLEG): container finished" podID="68f6ff1d-a119-4faf-89aa-fd9afc2d93f8" containerID="4965483e6277ff80bc871b75394d2f20f37e0b634fd01578030291d8d29bfe8a" exitCode=0 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.178127 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" event={"ID":"68f6ff1d-a119-4faf-89aa-fd9afc2d93f8","Type":"ContainerDied","Data":"4965483e6277ff80bc871b75394d2f20f37e0b634fd01578030291d8d29bfe8a"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.198565 4767 generic.go:334] "Generic (PLEG): container finished" podID="d24c721d-4968-477f-ba8e-23f6d31629d2" containerID="bb348c2b124362aa1be574374c3cd321b25ea76e59aee8aba55f432d9c45538f" exitCode=0 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.199709 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d24c721d-4968-477f-ba8e-23f6d31629d2","Type":"ContainerDied","Data":"bb348c2b124362aa1be574374c3cd321b25ea76e59aee8aba55f432d9c45538f"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.216666 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.222787 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.230617 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"77bdc76b5ac8583b5e07ed8ebbd2078302b0d1ecdca325886c659f034324896d"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.251713 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-config-operator_openshift-config-operator-7777fb866f-rphbj_3f692e30-f8c9-4823-bc82-7bf8abca0659/openshift-config-operator/0.log" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.273049 4767 generic.go:334] "Generic (PLEG): container finished" podID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerID="d1e085d042d9d8b5cec9ab40cfe208b9224e0fe81e8ca191bf191605cd7af4c7" exitCode=0 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.273201 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" event={"ID":"3f692e30-f8c9-4823-bc82-7bf8abca0659","Type":"ContainerDied","Data":"d1e085d042d9d8b5cec9ab40cfe208b9224e0fe81e8ca191bf191605cd7af4c7"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.273254 4767 scope.go:117] "RemoveContainer" containerID="0f8553bfc41dec31090eb663209a1e4657345284d2f6241ad40d2b406f04d846" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.301671 4767 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb" exitCode=0 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.301741 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"ad29f531e4d51f8b876dc324a3cc2b1a695944f3eecf342b6ed04626f2baf0eb"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.308894 4767 generic.go:334] "Generic (PLEG): container finished" podID="921c5e77-9858-4177-99a0-8cd3a4420d7d" containerID="204d75195a2f996056746fc3b56900ac3f45b8b912281c14865ffa9b0b1e339f" exitCode=1 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.311250 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" event={"ID":"921c5e77-9858-4177-99a0-8cd3a4420d7d","Type":"ContainerDied","Data":"204d75195a2f996056746fc3b56900ac3f45b8b912281c14865ffa9b0b1e339f"} Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.311789 4767 scope.go:117] "RemoveContainer" containerID="204d75195a2f996056746fc3b56900ac3f45b8b912281c14865ffa9b0b1e339f" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.313375 4767 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-82k6b container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.76:8443/healthz\": dial tcp 10.217.0.76:8443: connect: connection refused" start-of-body= Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.313412 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.76:8443/healthz\": dial tcp 10.217.0.76:8443: connect: connection refused" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.314037 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.314073 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.316454 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.316512 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.342500 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="6abe028c-416e-4978-bcc8-3a7b8d92624b" containerName="galera" containerID="cri-o://6abac3a757369ed7aa38da0ae38fc79753f588411284852e6d063714ed21a5fd" gracePeriod=29 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.379541 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" containerName="galera" containerID="cri-o://3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac" gracePeriod=18 Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.895500 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.895866 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.895518 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-rbpmk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.895917 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" podUID="3e0ef009-3626-492a-828e-be5f7285227c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 17 17:01:36 crc kubenswrapper[4767]: I0317 17:01:36.896903 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="7f03669c-8635-4948-b89c-0f2c7de39718" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 17:01:37 crc kubenswrapper[4767]: I0317 17:01:37.347408 4767 generic.go:334] "Generic (PLEG): container finished" podID="24b4b439-f309-4551-86d6-fc26f57fe754" containerID="44e01702cdd6851f3e6eac45406e1af0302eab03207b5ef699665002ceb62b55" exitCode=0 Mar 17 17:01:37 crc kubenswrapper[4767]: I0317 17:01:37.347590 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5jdf" event={"ID":"24b4b439-f309-4551-86d6-fc26f57fe754","Type":"ContainerDied","Data":"44e01702cdd6851f3e6eac45406e1af0302eab03207b5ef699665002ceb62b55"} Mar 17 17:01:37 crc kubenswrapper[4767]: I0317 17:01:37.356661 4767 generic.go:334] "Generic (PLEG): container finished" podID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerID="80cdce19b94103bc1982d3760e2748cdc19f5e1e69691384f05bd4ed1d8cf95e" exitCode=0 Mar 17 17:01:37 crc kubenswrapper[4767]: I0317 17:01:37.370607 4767 generic.go:334] "Generic (PLEG): container finished" podID="a7b5a1f5-da92-46dd-a5b0-5088e75346af" containerID="e0db98d18a0069ffcf1f782be7c5223dd0bbf104fd9ff358d8c6daac1bd5edef" exitCode=1 Mar 17 17:01:37 crc kubenswrapper[4767]: I0317 17:01:37.394745 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl4m8" event={"ID":"56a3dbff-bd98-4264-974a-5c098c6f1361","Type":"ContainerDied","Data":"80cdce19b94103bc1982d3760e2748cdc19f5e1e69691384f05bd4ed1d8cf95e"} Mar 17 17:01:37 crc kubenswrapper[4767]: I0317 17:01:37.395129 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" event={"ID":"a7b5a1f5-da92-46dd-a5b0-5088e75346af","Type":"ContainerDied","Data":"e0db98d18a0069ffcf1f782be7c5223dd0bbf104fd9ff358d8c6daac1bd5edef"} Mar 17 17:01:37 crc kubenswrapper[4767]: I0317 17:01:37.415199 4767 scope.go:117] "RemoveContainer" containerID="e0db98d18a0069ffcf1f782be7c5223dd0bbf104fd9ff358d8c6daac1bd5edef" Mar 17 17:01:37 crc kubenswrapper[4767]: I0317 17:01:37.831640 4767 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-82k6b container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.76:8443/healthz\": dial tcp 10.217.0.76:8443: connect: connection refused" start-of-body= Mar 17 17:01:37 crc kubenswrapper[4767]: I0317 17:01:37.832064 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.76:8443/healthz\": dial tcp 10.217.0.76:8443: connect: connection refused" Mar 17 17:01:37 crc kubenswrapper[4767]: I0317 17:01:37.833561 4767 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-82k6b container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.76:8443/healthz\": dial tcp 10.217.0.76:8443: connect: connection refused" start-of-body= Mar 17 17:01:37 crc kubenswrapper[4767]: I0317 17:01:37.833669 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" podUID="15e74b29-0b89-4e6a-84aa-94d9f5faa0e5" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.76:8443/healthz\": dial tcp 10.217.0.76:8443: connect: connection refused" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.055814 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.056164 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.443137 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" event={"ID":"6f682203-ee25-4b31-957b-26d8148feda0","Type":"ContainerStarted","Data":"f43523cc8d9123263473bac062d810ae489a0916efaa06cf09ab024d31a7d10f"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.444908 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.472225 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" event={"ID":"a333e040-ed72-4d18-95b1-14b57a92c895","Type":"ContainerStarted","Data":"21d8542388f3c6d94836c1e00ce12d3fe979ca9b8c6480f6262e1194e5ad61c4"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.472954 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.473012 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.474740 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.486470 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" event={"ID":"70fbe8aa-8647-4fe8-914c-0c05399cf46d","Type":"ContainerStarted","Data":"e0c7e20f3009faac390f6f73bc0810db0c48eb12e86a24db84c4e2b39ec27545"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.511821 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" event={"ID":"59ba3b25-1e95-41ad-921c-9ee4ec5e2c43","Type":"ContainerStarted","Data":"d799ae5a978d7dbdd630356c891a7662dea52dfcb357b7ec350dde1cd22841e1"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.512186 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.529239 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" event={"ID":"877f4eda-1ec7-4296-98df-b5ca7a7fa78a","Type":"ContainerStarted","Data":"f6a0cce18a10c62625237e3392b3d526c80a4757f05acf83f778624c9e0b4108"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.529747 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.539009 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" event={"ID":"1767bf72-c95d-486d-baa6-ca741742569b","Type":"ContainerStarted","Data":"63bdf8d2c32a72234e4e1c0e0a0ab775e28672b8b91a98dcbc5c34d849643331"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.539561 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.539935 4767 patch_prober.go:28] interesting pod/controller-manager-57bc79d78-25d96 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.539973 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podUID="1767bf72-c95d-486d-baa6-ca741742569b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.547289 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" event={"ID":"b836be84-e8b0-4e8d-8bac-cfc922bea7eb","Type":"ContainerStarted","Data":"3aae10d3c734e48fc5bd6cfd8e006280ce89aa94a18ebe2313f09149a8deb204"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.547663 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.547962 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.548028 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.552606 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" event={"ID":"3f692e30-f8c9-4823-bc82-7bf8abca0659","Type":"ContainerStarted","Data":"40dc07a253d9febfc8efed81f7046daed4bcfe885ea05335f7d622b7225ea335"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.552828 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.557460 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"81cba186a451b754cd05dcd8f54659545a79b32de2a766cbef5022976e4623e3"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.557968 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.561763 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" event={"ID":"921c5e77-9858-4177-99a0-8cd3a4420d7d","Type":"ContainerStarted","Data":"fad34314cbfeca070d20675502f744a281850b126ba8b15ec449e298422009d9"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.562542 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.567681 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" event={"ID":"e59ff072-51c2-4995-9f61-709f12a1393a","Type":"ContainerStarted","Data":"5b796506718844254dde871df541634795d49baf0ed05203f12e23891e049609"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.568559 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.578644 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" event={"ID":"a9f5b566-a215-4322-8d74-668dc2a93f3e","Type":"ContainerStarted","Data":"d22f6d2db417f23dc5e5d806090f1f5081473163b6759a35f02fa5c43280ba2b"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.580338 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.581253 4767 patch_prober.go:28] interesting pod/route-controller-manager-56d7bf8c9b-cs8b8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.581467 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" podUID="a9f5b566-a215-4322-8d74-668dc2a93f3e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.640128 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wj8xp" event={"ID":"e5da4ccd-4ecb-4929-8974-0a31d018204d","Type":"ContainerStarted","Data":"791e5af3ec4cdb48191f453e8f40fd97144bda8275fcc14c88fe318814245964"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.641873 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-wj8xp" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.663817 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" event={"ID":"07d5db0e-27a0-43ee-a8a7-54c27bf6f9ec","Type":"ContainerStarted","Data":"6e60a50f217d2416326ea2d83c98ba0c73a498582c1b1c1a78ca6f8a7b24dc62"} Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.665589 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 17:01:38 crc kubenswrapper[4767]: I0317 17:01:38.884261 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-kftt2" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.276334 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Mar 17 17:01:39 crc kubenswrapper[4767]: [+]has-synced ok Mar 17 17:01:39 crc kubenswrapper[4767]: [-]process-running failed: reason withheld Mar 17 17:01:39 crc kubenswrapper[4767]: healthz check failed Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.276709 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.476482 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="7f03669c-8635-4948-b89c-0f2c7de39718" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.498915 4767 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-lzjm6 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.499017 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" podUID="68f6ff1d-a119-4faf-89aa-fd9afc2d93f8" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.501444 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.501504 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.501526 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w7cg4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.501584 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" podUID="75f035cd-32d9-44a2-8d8f-3d5f40742b67" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.576983 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.577051 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.577104 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.577187 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.685920 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" event={"ID":"68f6ff1d-a119-4faf-89aa-fd9afc2d93f8","Type":"ContainerStarted","Data":"b68cf1617d67d43a4ea0d9f1ad523d9a57896c05f76fdb01f23c210cf6502d90"} Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.686100 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.705931 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d24c721d-4968-477f-ba8e-23f6d31629d2","Type":"ContainerStarted","Data":"bef6f215dbabec82bc0d2f9d5a1d7da229cddf82d189da1d7cf60b64bab09334"} Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.711968 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" event={"ID":"fda1c393-424a-4142-8570-5dc108f6d6f4","Type":"ContainerStarted","Data":"01e9e883e81b813b137d1434b0dfee5e2aaf316cdb98f03cab955f8ec0f6c9b7"} Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.712129 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.725785 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5jdf" event={"ID":"24b4b439-f309-4551-86d6-fc26f57fe754","Type":"ContainerStarted","Data":"a700d55e00187260c121ed7a6300e6394851eb6fcfa85b0a61eb9032655b41cf"} Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.744933 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mdlms" event={"ID":"a8829d5d-100e-4518-8863-db3ab9c7b30d","Type":"ContainerStarted","Data":"420264064f33020afc5f2fb4fe829b1e1d82577c291fc121f35c2a972e17ff3b"} Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.764223 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl4m8" event={"ID":"56a3dbff-bd98-4264-974a-5c098c6f1361","Type":"ContainerStarted","Data":"f3067e2134292fb78cd0e0c50fc4bfd58dea03350743321612315127cf59028a"} Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.769316 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" event={"ID":"904d9f3b-95f0-4e57-8d04-c2026227a4a6","Type":"ContainerStarted","Data":"72585b7f5fd5772ebdb3902734d7aab319772964badfd8db2ba2a270ad56409b"} Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.770412 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.776328 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" event={"ID":"a7b5a1f5-da92-46dd-a5b0-5088e75346af","Type":"ContainerStarted","Data":"608257b0390a9552487ed151349ebad537fb9dd365e9d490742c3e7e6f94ac8b"} Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.777128 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.782916 4767 patch_prober.go:28] interesting pod/controller-manager-57bc79d78-25d96 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.782981 4767 patch_prober.go:28] interesting pod/route-controller-manager-56d7bf8c9b-cs8b8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.783006 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podUID="1767bf72-c95d-486d-baa6-ca741742569b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.783048 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" podUID="a9f5b566-a215-4322-8d74-668dc2a93f3e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.782916 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.783082 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.784727 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.784811 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 17 17:01:39 crc kubenswrapper[4767]: I0317 17:01:39.795658 4767 trace.go:236] Trace[1103100637]: "Calculate volume metrics of glance for pod openstack/glance-default-internal-api-0" (17-Mar-2026 17:01:38.535) (total time: 1260ms): Mar 17 17:01:39 crc kubenswrapper[4767]: Trace[1103100637]: [1.260203601s] [1.260203601s] END Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.237514 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.237897 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.238871 4767 patch_prober.go:28] interesting pod/controller-manager-57bc79d78-25d96 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.238928 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" podUID="1767bf72-c95d-486d-baa6-ca741742569b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.239187 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8rfjw container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.239207 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" podUID="b836be84-e8b0-4e8d-8bac-cfc922bea7eb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.243612 4767 patch_prober.go:28] interesting pod/route-controller-manager-56d7bf8c9b-cs8b8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.243704 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" podUID="a9f5b566-a215-4322-8d74-668dc2a93f3e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.245661 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.267413 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.452301 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.556077 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.556439 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.885559 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5444994796-ztnfb_a5cf3c29-3814-4d17-876a-9ca3e44a9400/router/0.log" Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.885639 4767 generic.go:334] "Generic (PLEG): container finished" podID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerID="3dbc65c86754010059515f8fc9873782ca67f8577939e37319a0dcd29bf71113" exitCode=137 Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.885803 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ztnfb" event={"ID":"a5cf3c29-3814-4d17-876a-9ca3e44a9400","Type":"ContainerDied","Data":"3dbc65c86754010059515f8fc9873782ca67f8577939e37319a0dcd29bf71113"} Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.903845 4767 generic.go:334] "Generic (PLEG): container finished" podID="6abe028c-416e-4978-bcc8-3a7b8d92624b" containerID="6abac3a757369ed7aa38da0ae38fc79753f588411284852e6d063714ed21a5fd" exitCode=0 Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.904093 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6abe028c-416e-4978-bcc8-3a7b8d92624b","Type":"ContainerDied","Data":"6abac3a757369ed7aa38da0ae38fc79753f588411284852e6d063714ed21a5fd"} Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.909250 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-6r96q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 17 17:01:40 crc kubenswrapper[4767]: I0317 17:01:40.909340 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" podUID="a333e040-ed72-4d18-95b1-14b57a92c895" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 17 17:01:41 crc kubenswrapper[4767]: I0317 17:01:41.058835 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 17:01:41 crc kubenswrapper[4767]: I0317 17:01:41.058967 4767 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rphbj container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 17 17:01:41 crc kubenswrapper[4767]: I0317 17:01:41.059383 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 17:01:41 crc kubenswrapper[4767]: I0317 17:01:41.059213 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" podUID="3f692e30-f8c9-4823-bc82-7bf8abca0659" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 17 17:01:41 crc kubenswrapper[4767]: I0317 17:01:41.078381 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-d7dccc75b-zlss6" Mar 17 17:01:41 crc kubenswrapper[4767]: I0317 17:01:41.918649 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5444994796-ztnfb_a5cf3c29-3814-4d17-876a-9ca3e44a9400/router/0.log" Mar 17 17:01:41 crc kubenswrapper[4767]: I0317 17:01:41.919278 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ztnfb" event={"ID":"a5cf3c29-3814-4d17-876a-9ca3e44a9400","Type":"ContainerStarted","Data":"004b45e3b496f13755e8f042eef7dc054f12c5436ce1e7b27fc0538f7ec2985f"} Mar 17 17:01:41 crc kubenswrapper[4767]: I0317 17:01:41.927755 4767 generic.go:334] "Generic (PLEG): container finished" podID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" containerID="3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac" exitCode=0 Mar 17 17:01:41 crc kubenswrapper[4767]: I0317 17:01:41.927858 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271","Type":"ContainerDied","Data":"3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac"} Mar 17 17:01:42 crc kubenswrapper[4767]: I0317 17:01:42.147719 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:42 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:42 crc kubenswrapper[4767]: > Mar 17 17:01:42 crc kubenswrapper[4767]: I0317 17:01:42.265129 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 17:01:42 crc kubenswrapper[4767]: I0317 17:01:42.266842 4767 patch_prober.go:28] interesting pod/router-default-5444994796-ztnfb container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 17 17:01:42 crc kubenswrapper[4767]: I0317 17:01:42.273553 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ztnfb" podUID="a5cf3c29-3814-4d17-876a-9ca3e44a9400" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 17 17:01:42 crc kubenswrapper[4767]: I0317 17:01:42.334933 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 17:01:42 crc kubenswrapper[4767]: I0317 17:01:42.409889 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-89866dfb6-fswnc" Mar 17 17:01:42 crc kubenswrapper[4767]: E0317 17:01:42.490593 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac is running failed: container process not found" containerID="3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 17 17:01:42 crc kubenswrapper[4767]: E0317 17:01:42.491141 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac is running failed: container process not found" containerID="3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 17 17:01:42 crc kubenswrapper[4767]: E0317 17:01:42.491423 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac is running failed: container process not found" containerID="3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 17 17:01:42 crc kubenswrapper[4767]: E0317 17:01:42.491454 4767 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3d1989ebe809ef5069e7525f3bdbd5b895624b712cf477204a7b4b3c21a395ac is running failed: container process not found" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271" containerName="galera" Mar 17 17:01:42 crc kubenswrapper[4767]: I0317 17:01:42.843610 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="7f03669c-8635-4948-b89c-0f2c7de39718" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 17:01:42 crc kubenswrapper[4767]: I0317 17:01:42.844338 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 17 17:01:42 crc kubenswrapper[4767]: I0317 17:01:42.846244 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cinder-scheduler" containerStatusID={"Type":"cri-o","ID":"72c392e8dd7e292a16bc4cfcbd28413bffec2553e5bfca74eaf17003de7080e8"} pod="openstack/cinder-scheduler-0" containerMessage="Container cinder-scheduler failed liveness probe, will be restarted" Mar 17 17:01:42 crc kubenswrapper[4767]: I0317 17:01:42.846452 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7f03669c-8635-4948-b89c-0f2c7de39718" containerName="cinder-scheduler" containerID="cri-o://72c392e8dd7e292a16bc4cfcbd28413bffec2553e5bfca74eaf17003de7080e8" gracePeriod=30 Mar 17 17:01:43 crc kubenswrapper[4767]: I0317 17:01:43.010726 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b9b5f4cd-bbe9-4c5d-8d0e-83b7cf4c8271","Type":"ContainerStarted","Data":"c0bd4eba2468a98bf4091d24f578fcfd3cdc51b129103325bdbb4d11481e6d12"} Mar 17 17:01:43 crc kubenswrapper[4767]: I0317 17:01:43.017313 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6abe028c-416e-4978-bcc8-3a7b8d92624b","Type":"ContainerStarted","Data":"711687a84c32d0e0ef6ced0643c08d5aa01662a6b12edabb37bd73d4c7e2f6f8"} Mar 17 17:01:43 crc kubenswrapper[4767]: I0317 17:01:43.120567 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 17:01:43 crc kubenswrapper[4767]: I0317 17:01:43.283694 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 17:01:43 crc kubenswrapper[4767]: I0317 17:01:43.582241 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-24wkj" Mar 17 17:01:43 crc kubenswrapper[4767]: I0317 17:01:43.591046 4767 patch_prober.go:28] interesting pod/loki-operator-controller-manager-5dcc96cb6-rv7xm container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.50:8081/readyz\": dial tcp 10.217.0.50:8081: connect: connection refused" start-of-body= Mar 17 17:01:43 crc kubenswrapper[4767]: I0317 17:01:43.591101 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" podUID="6f682203-ee25-4b31-957b-26d8148feda0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.50:8081/readyz\": dial tcp 10.217.0.50:8081: connect: connection refused" Mar 17 17:01:43 crc kubenswrapper[4767]: I0317 17:01:43.929778 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 17 17:01:43 crc kubenswrapper[4767]: I0317 17:01:43.929996 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 17 17:01:44 crc kubenswrapper[4767]: I0317 17:01:44.087914 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 17:01:44 crc kubenswrapper[4767]: I0317 17:01:44.102119 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-x6vcv" Mar 17 17:01:44 crc kubenswrapper[4767]: I0317 17:01:44.136505 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ztnfb" Mar 17 17:01:44 crc kubenswrapper[4767]: I0317 17:01:44.179699 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 17:01:44 crc kubenswrapper[4767]: I0317 17:01:44.179774 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 17:01:44 crc kubenswrapper[4767]: I0317 17:01:44.202632 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rphbj" Mar 17 17:01:44 crc kubenswrapper[4767]: I0317 17:01:44.560477 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-r8z96" Mar 17 17:01:44 crc kubenswrapper[4767]: I0317 17:01:44.810644 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-xms5z" Mar 17 17:01:44 crc kubenswrapper[4767]: I0317 17:01:44.853085 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8f8r" Mar 17 17:01:45 crc kubenswrapper[4767]: I0317 17:01:45.293398 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-sl4m8" podUID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:45 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:45 crc kubenswrapper[4767]: > Mar 17 17:01:46 crc kubenswrapper[4767]: I0317 17:01:46.120105 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-wjz85" Mar 17 17:01:46 crc kubenswrapper[4767]: I0317 17:01:46.901187 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-rbpmk" Mar 17 17:01:47 crc kubenswrapper[4767]: I0317 17:01:47.869884 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-82k6b" Mar 17 17:01:49 crc kubenswrapper[4767]: I0317 17:01:49.509494 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w7cg4" Mar 17 17:01:49 crc kubenswrapper[4767]: I0317 17:01:49.582562 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6r96q" Mar 17 17:01:49 crc kubenswrapper[4767]: I0317 17:01:49.984420 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 17:01:49 crc kubenswrapper[4767]: I0317 17:01:49.990993 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57bc79d78-25d96" Mar 17 17:01:50 crc kubenswrapper[4767]: I0317 17:01:49.999393 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-56d7bf8c9b-cs8b8" Mar 17 17:01:50 crc kubenswrapper[4767]: I0317 17:01:50.157273 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8rfjw" Mar 17 17:01:50 crc kubenswrapper[4767]: I0317 17:01:50.229819 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7f03669c-8635-4948-b89c-0f2c7de39718","Type":"ContainerDied","Data":"72c392e8dd7e292a16bc4cfcbd28413bffec2553e5bfca74eaf17003de7080e8"} Mar 17 17:01:50 crc kubenswrapper[4767]: I0317 17:01:50.229897 4767 generic.go:334] "Generic (PLEG): container finished" podID="7f03669c-8635-4948-b89c-0f2c7de39718" containerID="72c392e8dd7e292a16bc4cfcbd28413bffec2553e5bfca74eaf17003de7080e8" exitCode=0 Mar 17 17:01:50 crc kubenswrapper[4767]: I0317 17:01:50.463482 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-pjldm" Mar 17 17:01:51 crc kubenswrapper[4767]: I0317 17:01:51.489616 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5974d9c54d-9rs7d" Mar 17 17:01:51 crc kubenswrapper[4767]: I0317 17:01:51.688557 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:51 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:51 crc kubenswrapper[4767]: > Mar 17 17:01:52 crc kubenswrapper[4767]: I0317 17:01:52.274749 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7f03669c-8635-4948-b89c-0f2c7de39718","Type":"ContainerStarted","Data":"05c776b911380661fefcaba2115404e9e2aaf7424acbc841059a2672443631d7"} Mar 17 17:01:52 crc kubenswrapper[4767]: I0317 17:01:52.424159 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-wj8xp" Mar 17 17:01:52 crc kubenswrapper[4767]: I0317 17:01:52.440211 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-khzr7" Mar 17 17:01:52 crc kubenswrapper[4767]: I0317 17:01:52.490073 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 17 17:01:52 crc kubenswrapper[4767]: I0317 17:01:52.490145 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 17 17:01:53 crc kubenswrapper[4767]: I0317 17:01:53.363272 4767 generic.go:334] "Generic (PLEG): container finished" podID="909f6b35-823c-4b1b-ba3b-6db1a4f6b444" containerID="c8b30c60aed4f751fdd0bc2ff9a5fc39033ebe3dc1a35d61b2a2e94fd217f757" exitCode=1 Mar 17 17:01:53 crc kubenswrapper[4767]: I0317 17:01:53.379777 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"909f6b35-823c-4b1b-ba3b-6db1a4f6b444","Type":"ContainerDied","Data":"c8b30c60aed4f751fdd0bc2ff9a5fc39033ebe3dc1a35d61b2a2e94fd217f757"} Mar 17 17:01:53 crc kubenswrapper[4767]: I0317 17:01:53.592695 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-5dcc96cb6-rv7xm" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.241843 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-sl4m8" podUID="56a3dbff-bd98-4264-974a-5c098c6f1361" containerName="registry-server" probeResult="failure" output=< Mar 17 17:01:55 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:01:55 crc kubenswrapper[4767]: > Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.425254 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"909f6b35-823c-4b1b-ba3b-6db1a4f6b444","Type":"ContainerDied","Data":"e3039e512befc9505271561e6ed0eeab8562ce4f3543358b67aa0d2b146eb99c"} Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.425307 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3039e512befc9505271561e6ed0eeab8562ce4f3543358b67aa0d2b146eb99c" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.498937 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.658267 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-config-data\") pod \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.658348 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ssh-key\") pod \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.658387 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config\") pod \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.658417 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ca-certs\") pod \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.658581 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-temporary\") pod \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.658610 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config-secret\") pod \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.658643 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4zw7\" (UniqueName: \"kubernetes.io/projected/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-kube-api-access-k4zw7\") pod \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.658820 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-workdir\") pod \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.658849 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\" (UID: \"909f6b35-823c-4b1b-ba3b-6db1a4f6b444\") " Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.661525 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "909f6b35-823c-4b1b-ba3b-6db1a4f6b444" (UID: "909f6b35-823c-4b1b-ba3b-6db1a4f6b444"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.663021 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-config-data" (OuterVolumeSpecName: "config-data") pod "909f6b35-823c-4b1b-ba3b-6db1a4f6b444" (UID: "909f6b35-823c-4b1b-ba3b-6db1a4f6b444"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.668667 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "909f6b35-823c-4b1b-ba3b-6db1a4f6b444" (UID: "909f6b35-823c-4b1b-ba3b-6db1a4f6b444"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.670297 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "909f6b35-823c-4b1b-ba3b-6db1a4f6b444" (UID: "909f6b35-823c-4b1b-ba3b-6db1a4f6b444"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.677375 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-kube-api-access-k4zw7" (OuterVolumeSpecName: "kube-api-access-k4zw7") pod "909f6b35-823c-4b1b-ba3b-6db1a4f6b444" (UID: "909f6b35-823c-4b1b-ba3b-6db1a4f6b444"). InnerVolumeSpecName "kube-api-access-k4zw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.707233 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "909f6b35-823c-4b1b-ba3b-6db1a4f6b444" (UID: "909f6b35-823c-4b1b-ba3b-6db1a4f6b444"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.723000 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "909f6b35-823c-4b1b-ba3b-6db1a4f6b444" (UID: "909f6b35-823c-4b1b-ba3b-6db1a4f6b444"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.734350 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "909f6b35-823c-4b1b-ba3b-6db1a4f6b444" (UID: "909f6b35-823c-4b1b-ba3b-6db1a4f6b444"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.751565 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "909f6b35-823c-4b1b-ba3b-6db1a4f6b444" (UID: "909f6b35-823c-4b1b-ba3b-6db1a4f6b444"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.762608 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.762652 4767 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ssh-key\") on node \"crc\" DevicePath \"\"" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.762666 4767 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.762679 4767 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-ca-certs\") on node \"crc\" DevicePath \"\"" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.762691 4767 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.762702 4767 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.762713 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4zw7\" (UniqueName: \"kubernetes.io/projected/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-kube-api-access-k4zw7\") on node \"crc\" DevicePath \"\"" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.762722 4767 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/909f6b35-823c-4b1b-ba3b-6db1a4f6b444-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.763114 4767 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.794014 4767 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 17 17:01:55 crc kubenswrapper[4767]: I0317 17:01:55.866100 4767 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 17 17:01:56 crc kubenswrapper[4767]: I0317 17:01:56.207619 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" podUID="2321f9fb-78c8-401b-8fc2-a5f82d42a226" containerName="oauth-openshift" containerID="cri-o://eaff0e00c90c96aca126c5718dbe75b4480b9cf9712683c7700af0e35030c21d" gracePeriod=15 Mar 17 17:01:56 crc kubenswrapper[4767]: I0317 17:01:56.446059 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 17 17:01:56 crc kubenswrapper[4767]: I0317 17:01:56.472850 4767 generic.go:334] "Generic (PLEG): container finished" podID="2321f9fb-78c8-401b-8fc2-a5f82d42a226" containerID="eaff0e00c90c96aca126c5718dbe75b4480b9cf9712683c7700af0e35030c21d" exitCode=0 Mar 17 17:01:56 crc kubenswrapper[4767]: I0317 17:01:56.472984 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 17 17:01:56 crc kubenswrapper[4767]: I0317 17:01:56.479735 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" event={"ID":"2321f9fb-78c8-401b-8fc2-a5f82d42a226","Type":"ContainerDied","Data":"eaff0e00c90c96aca126c5718dbe75b4480b9cf9712683c7700af0e35030c21d"} Mar 17 17:01:56 crc kubenswrapper[4767]: I0317 17:01:56.488690 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7f03669c-8635-4948-b89c-0f2c7de39718" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 17:01:57 crc kubenswrapper[4767]: I0317 17:01:57.490003 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" event={"ID":"2321f9fb-78c8-401b-8fc2-a5f82d42a226","Type":"ContainerStarted","Data":"37dfd4fc37eb66bde3f0320d630d2e3f20e65682a3d60a0bcb36652045ea7d45"} Mar 17 17:01:57 crc kubenswrapper[4767]: I0317 17:01:57.490434 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 17:01:57 crc kubenswrapper[4767]: I0317 17:01:57.491045 4767 patch_prober.go:28] interesting pod/oauth-openshift-5cc5b65bd-6nll7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" start-of-body= Mar 17 17:01:57 crc kubenswrapper[4767]: I0317 17:01:57.491106 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" podUID="2321f9fb-78c8-401b-8fc2-a5f82d42a226" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" Mar 17 17:01:57 crc kubenswrapper[4767]: I0317 17:01:57.923778 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 17 17:01:57 crc kubenswrapper[4767]: E0317 17:01:57.935580 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909f6b35-823c-4b1b-ba3b-6db1a4f6b444" containerName="tempest-tests-tempest-tests-runner" Mar 17 17:01:57 crc kubenswrapper[4767]: I0317 17:01:57.935640 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="909f6b35-823c-4b1b-ba3b-6db1a4f6b444" containerName="tempest-tests-tempest-tests-runner" Mar 17 17:01:57 crc kubenswrapper[4767]: I0317 17:01:57.936652 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="909f6b35-823c-4b1b-ba3b-6db1a4f6b444" containerName="tempest-tests-tempest-tests-runner" Mar 17 17:01:57 crc kubenswrapper[4767]: I0317 17:01:57.939357 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 17:01:57 crc kubenswrapper[4767]: I0317 17:01:57.949959 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-c8ndj" Mar 17 17:01:57 crc kubenswrapper[4767]: I0317 17:01:57.951028 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 17 17:01:58 crc kubenswrapper[4767]: I0317 17:01:58.059819 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lr65\" (UniqueName: \"kubernetes.io/projected/72354195-0d3e-4b5b-a026-a90f5e5f0dd3-kube-api-access-8lr65\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"72354195-0d3e-4b5b-a026-a90f5e5f0dd3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 17:01:58 crc kubenswrapper[4767]: I0317 17:01:58.061163 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"72354195-0d3e-4b5b-a026-a90f5e5f0dd3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 17:01:58 crc kubenswrapper[4767]: I0317 17:01:58.163548 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"72354195-0d3e-4b5b-a026-a90f5e5f0dd3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 17:01:58 crc kubenswrapper[4767]: I0317 17:01:58.163668 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lr65\" (UniqueName: \"kubernetes.io/projected/72354195-0d3e-4b5b-a026-a90f5e5f0dd3-kube-api-access-8lr65\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"72354195-0d3e-4b5b-a026-a90f5e5f0dd3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 17:01:58 crc kubenswrapper[4767]: I0317 17:01:58.164191 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"72354195-0d3e-4b5b-a026-a90f5e5f0dd3\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 17:01:58 crc kubenswrapper[4767]: I0317 17:01:58.193776 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lr65\" (UniqueName: \"kubernetes.io/projected/72354195-0d3e-4b5b-a026-a90f5e5f0dd3-kube-api-access-8lr65\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"72354195-0d3e-4b5b-a026-a90f5e5f0dd3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 17:01:58 crc kubenswrapper[4767]: I0317 17:01:58.204398 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"72354195-0d3e-4b5b-a026-a90f5e5f0dd3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 17:01:58 crc kubenswrapper[4767]: I0317 17:01:58.301072 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 17:01:58 crc kubenswrapper[4767]: I0317 17:01:58.508251 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5cc5b65bd-6nll7" Mar 17 17:01:59 crc kubenswrapper[4767]: I0317 17:01:59.044929 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 17 17:01:59 crc kubenswrapper[4767]: I0317 17:01:59.521203 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"72354195-0d3e-4b5b-a026-a90f5e5f0dd3","Type":"ContainerStarted","Data":"3be0bcf6926824369871b850a44404e4f2244737e27289035e42b6019863d563"} Mar 17 17:02:01 crc kubenswrapper[4767]: I0317 17:02:01.589471 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"72354195-0d3e-4b5b-a026-a90f5e5f0dd3","Type":"ContainerStarted","Data":"a54b04ceca51e85ef30ebe5b7aaa760d922279a6bfe0048399f4a2b43ebb7fa3"} Mar 17 17:02:01 crc kubenswrapper[4767]: I0317 17:02:01.596480 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7f03669c-8635-4948-b89c-0f2c7de39718" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 17:02:01 crc kubenswrapper[4767]: I0317 17:02:01.643616 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=3.386947277 podStartE2EDuration="4.639998945s" podCreationTimestamp="2026-03-17 17:01:57 +0000 UTC" firstStartedPulling="2026-03-17 17:01:59.062908592 +0000 UTC m=+5110.476224649" lastFinishedPulling="2026-03-17 17:02:00.31596027 +0000 UTC m=+5111.729276317" observedRunningTime="2026-03-17 17:02:01.632108801 +0000 UTC m=+5113.045424858" watchObservedRunningTime="2026-03-17 17:02:01.639998945 +0000 UTC m=+5113.053314992" Mar 17 17:02:02 crc kubenswrapper[4767]: I0317 17:02:02.105282 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output=< Mar 17 17:02:02 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:02:02 crc kubenswrapper[4767]: > Mar 17 17:02:02 crc kubenswrapper[4767]: I0317 17:02:02.792862 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562782-x6clm"] Mar 17 17:02:02 crc kubenswrapper[4767]: I0317 17:02:02.802295 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562782-x6clm" Mar 17 17:02:02 crc kubenswrapper[4767]: I0317 17:02:02.809964 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 17:02:02 crc kubenswrapper[4767]: I0317 17:02:02.811093 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 17:02:02 crc kubenswrapper[4767]: I0317 17:02:02.812048 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562782-x6clm"] Mar 17 17:02:02 crc kubenswrapper[4767]: I0317 17:02:02.830954 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 17:02:03 crc kubenswrapper[4767]: I0317 17:02:03.055062 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhnn9\" (UniqueName: \"kubernetes.io/projected/b00fe51b-6938-436c-ae5b-f961f538af6b-kube-api-access-dhnn9\") pod \"auto-csr-approver-29562782-x6clm\" (UID: \"b00fe51b-6938-436c-ae5b-f961f538af6b\") " pod="openshift-infra/auto-csr-approver-29562782-x6clm" Mar 17 17:02:03 crc kubenswrapper[4767]: I0317 17:02:03.158533 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhnn9\" (UniqueName: \"kubernetes.io/projected/b00fe51b-6938-436c-ae5b-f961f538af6b-kube-api-access-dhnn9\") pod \"auto-csr-approver-29562782-x6clm\" (UID: \"b00fe51b-6938-436c-ae5b-f961f538af6b\") " pod="openshift-infra/auto-csr-approver-29562782-x6clm" Mar 17 17:02:03 crc kubenswrapper[4767]: I0317 17:02:03.703638 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhnn9\" (UniqueName: \"kubernetes.io/projected/b00fe51b-6938-436c-ae5b-f961f538af6b-kube-api-access-dhnn9\") pod \"auto-csr-approver-29562782-x6clm\" (UID: \"b00fe51b-6938-436c-ae5b-f961f538af6b\") " pod="openshift-infra/auto-csr-approver-29562782-x6clm" Mar 17 17:02:03 crc kubenswrapper[4767]: I0317 17:02:03.740315 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562782-x6clm" Mar 17 17:02:04 crc kubenswrapper[4767]: I0317 17:02:04.207899 4767 patch_prober.go:28] interesting pod/machine-config-daemon-lhrtr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 17:02:04 crc kubenswrapper[4767]: I0317 17:02:04.208219 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 17:02:04 crc kubenswrapper[4767]: I0317 17:02:04.208278 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" Mar 17 17:02:04 crc kubenswrapper[4767]: I0317 17:02:04.212633 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e"} pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 17:02:04 crc kubenswrapper[4767]: I0317 17:02:04.220594 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerName="machine-config-daemon" containerID="cri-o://1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" gracePeriod=600 Mar 17 17:02:04 crc kubenswrapper[4767]: E0317 17:02:04.407905 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:02:04 crc kubenswrapper[4767]: I0317 17:02:04.413999 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 17:02:04 crc kubenswrapper[4767]: I0317 17:02:04.645999 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sl4m8" Mar 17 17:02:04 crc kubenswrapper[4767]: I0317 17:02:04.739808 4767 generic.go:334] "Generic (PLEG): container finished" podID="d680dc75-619e-41ae-90cb-2f58898a33ac" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" exitCode=0 Mar 17 17:02:04 crc kubenswrapper[4767]: I0317 17:02:04.739975 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerDied","Data":"1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e"} Mar 17 17:02:04 crc kubenswrapper[4767]: I0317 17:02:04.747759 4767 scope.go:117] "RemoveContainer" containerID="c4883088f1ca02db880affb212da5ef91b29d53adeb7d812b4a4c5c2a0d946f2" Mar 17 17:02:04 crc kubenswrapper[4767]: I0317 17:02:04.748039 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:02:04 crc kubenswrapper[4767]: E0317 17:02:04.749096 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:02:05 crc kubenswrapper[4767]: I0317 17:02:05.140015 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562782-x6clm"] Mar 17 17:02:05 crc kubenswrapper[4767]: I0317 17:02:05.762995 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562782-x6clm" event={"ID":"b00fe51b-6938-436c-ae5b-f961f538af6b","Type":"ContainerStarted","Data":"5ae3bf804928dba8204b35c07fc94756977003ee52e5f98c8f1f47d46ca2ce11"} Mar 17 17:02:06 crc kubenswrapper[4767]: I0317 17:02:06.618066 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7f03669c-8635-4948-b89c-0f2c7de39718" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 17:02:07 crc kubenswrapper[4767]: I0317 17:02:07.479667 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 17 17:02:07 crc kubenswrapper[4767]: I0317 17:02:07.627802 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 17 17:02:07 crc kubenswrapper[4767]: I0317 17:02:07.866144 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 17 17:02:08 crc kubenswrapper[4767]: I0317 17:02:08.026379 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 17 17:02:08 crc kubenswrapper[4767]: I0317 17:02:08.834220 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562782-x6clm" event={"ID":"b00fe51b-6938-436c-ae5b-f961f538af6b","Type":"ContainerStarted","Data":"b2be0e35d3b97dfb8a377e0fca0f86e5a699664ab31654091483c1e3634bdcb8"} Mar 17 17:02:08 crc kubenswrapper[4767]: I0317 17:02:08.857548 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562782-x6clm" podStartSLOduration=5.227296612 podStartE2EDuration="6.857515813s" podCreationTimestamp="2026-03-17 17:02:02 +0000 UTC" firstStartedPulling="2026-03-17 17:02:05.202491345 +0000 UTC m=+5116.615807392" lastFinishedPulling="2026-03-17 17:02:06.832710546 +0000 UTC m=+5118.246026593" observedRunningTime="2026-03-17 17:02:08.849415923 +0000 UTC m=+5120.262731970" watchObservedRunningTime="2026-03-17 17:02:08.857515813 +0000 UTC m=+5120.270831860" Mar 17 17:02:09 crc kubenswrapper[4767]: I0317 17:02:09.604401 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzjm6" Mar 17 17:02:10 crc kubenswrapper[4767]: I0317 17:02:10.861521 4767 generic.go:334] "Generic (PLEG): container finished" podID="d38f4981-a828-4985-9162-9597fd4dfc6d" containerID="e52c8a922ac5726ae258ab86cf09377d32f8e065ff7f13bc06511afd70ebdfb0" exitCode=0 Mar 17 17:02:10 crc kubenswrapper[4767]: I0317 17:02:10.861594 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29562781-d47zt" event={"ID":"d38f4981-a828-4985-9162-9597fd4dfc6d","Type":"ContainerDied","Data":"e52c8a922ac5726ae258ab86cf09377d32f8e065ff7f13bc06511afd70ebdfb0"} Mar 17 17:02:10 crc kubenswrapper[4767]: I0317 17:02:10.864506 4767 generic.go:334] "Generic (PLEG): container finished" podID="b00fe51b-6938-436c-ae5b-f961f538af6b" containerID="b2be0e35d3b97dfb8a377e0fca0f86e5a699664ab31654091483c1e3634bdcb8" exitCode=0 Mar 17 17:02:10 crc kubenswrapper[4767]: I0317 17:02:10.864553 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562782-x6clm" event={"ID":"b00fe51b-6938-436c-ae5b-f961f538af6b","Type":"ContainerDied","Data":"b2be0e35d3b97dfb8a377e0fca0f86e5a699664ab31654091483c1e3634bdcb8"} Mar 17 17:02:11 crc kubenswrapper[4767]: I0317 17:02:11.192508 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5c5f94f4bc-xqf4r" Mar 17 17:02:11 crc kubenswrapper[4767]: I0317 17:02:11.499615 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 17 17:02:11 crc kubenswrapper[4767]: I0317 17:02:11.636215 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output=< Mar 17 17:02:11 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:02:11 crc kubenswrapper[4767]: > Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.803128 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562782-x6clm" Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.813005 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.895347 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29562781-d47zt" event={"ID":"d38f4981-a828-4985-9162-9597fd4dfc6d","Type":"ContainerDied","Data":"eb4e7a2dd2b9933ec39645e7e0f00e2d45b522e1e2f64e225e3e74274c1fefe6"} Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.895397 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb4e7a2dd2b9933ec39645e7e0f00e2d45b522e1e2f64e225e3e74274c1fefe6" Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.895407 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29562781-d47zt" Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.898391 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562782-x6clm" event={"ID":"b00fe51b-6938-436c-ae5b-f961f538af6b","Type":"ContainerDied","Data":"5ae3bf804928dba8204b35c07fc94756977003ee52e5f98c8f1f47d46ca2ce11"} Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.898434 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ae3bf804928dba8204b35c07fc94756977003ee52e5f98c8f1f47d46ca2ce11" Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.898490 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562782-x6clm" Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.976966 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-config-data\") pod \"d38f4981-a828-4985-9162-9597fd4dfc6d\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.977038 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-fernet-keys\") pod \"d38f4981-a828-4985-9162-9597fd4dfc6d\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.977304 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhnn9\" (UniqueName: \"kubernetes.io/projected/b00fe51b-6938-436c-ae5b-f961f538af6b-kube-api-access-dhnn9\") pod \"b00fe51b-6938-436c-ae5b-f961f538af6b\" (UID: \"b00fe51b-6938-436c-ae5b-f961f538af6b\") " Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.977382 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-combined-ca-bundle\") pod \"d38f4981-a828-4985-9162-9597fd4dfc6d\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " Mar 17 17:02:12 crc kubenswrapper[4767]: I0317 17:02:12.977448 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h9fc\" (UniqueName: \"kubernetes.io/projected/d38f4981-a828-4985-9162-9597fd4dfc6d-kube-api-access-9h9fc\") pod \"d38f4981-a828-4985-9162-9597fd4dfc6d\" (UID: \"d38f4981-a828-4985-9162-9597fd4dfc6d\") " Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.023400 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d38f4981-a828-4985-9162-9597fd4dfc6d" (UID: "d38f4981-a828-4985-9162-9597fd4dfc6d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.055054 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b00fe51b-6938-436c-ae5b-f961f538af6b-kube-api-access-dhnn9" (OuterVolumeSpecName: "kube-api-access-dhnn9") pod "b00fe51b-6938-436c-ae5b-f961f538af6b" (UID: "b00fe51b-6938-436c-ae5b-f961f538af6b"). InnerVolumeSpecName "kube-api-access-dhnn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.073736 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38f4981-a828-4985-9162-9597fd4dfc6d-kube-api-access-9h9fc" (OuterVolumeSpecName: "kube-api-access-9h9fc") pod "d38f4981-a828-4985-9162-9597fd4dfc6d" (UID: "d38f4981-a828-4985-9162-9597fd4dfc6d"). InnerVolumeSpecName "kube-api-access-9h9fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.084758 4767 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.084810 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhnn9\" (UniqueName: \"kubernetes.io/projected/b00fe51b-6938-436c-ae5b-f961f538af6b-kube-api-access-dhnn9\") on node \"crc\" DevicePath \"\"" Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.084825 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h9fc\" (UniqueName: \"kubernetes.io/projected/d38f4981-a828-4985-9162-9597fd4dfc6d-kube-api-access-9h9fc\") on node \"crc\" DevicePath \"\"" Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.116949 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562776-dt5ql"] Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.129984 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562776-dt5ql"] Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.133637 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d38f4981-a828-4985-9162-9597fd4dfc6d" (UID: "d38f4981-a828-4985-9162-9597fd4dfc6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.180848 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-config-data" (OuterVolumeSpecName: "config-data") pod "d38f4981-a828-4985-9162-9597fd4dfc6d" (UID: "d38f4981-a828-4985-9162-9597fd4dfc6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.188905 4767 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.188942 4767 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38f4981-a828-4985-9162-9597fd4dfc6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 17:02:13 crc kubenswrapper[4767]: I0317 17:02:13.374334 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8bcb562-f3bf-4856-bbeb-b67ed48d147f" path="/var/lib/kubelet/pods/a8bcb562-f3bf-4856-bbeb-b67ed48d147f/volumes" Mar 17 17:02:15 crc kubenswrapper[4767]: I0317 17:02:15.360885 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:02:15 crc kubenswrapper[4767]: E0317 17:02:15.362471 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:02:21 crc kubenswrapper[4767]: I0317 17:02:21.630447 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output=< Mar 17 17:02:21 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:02:21 crc kubenswrapper[4767]: > Mar 17 17:02:26 crc kubenswrapper[4767]: I0317 17:02:26.355759 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:02:26 crc kubenswrapper[4767]: E0317 17:02:26.356660 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:02:30 crc kubenswrapper[4767]: I0317 17:02:30.477394 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 17:02:31 crc kubenswrapper[4767]: I0317 17:02:31.677344 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v5jdf" podUID="24b4b439-f309-4551-86d6-fc26f57fe754" containerName="registry-server" probeResult="failure" output=< Mar 17 17:02:31 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:02:31 crc kubenswrapper[4767]: > Mar 17 17:02:37 crc kubenswrapper[4767]: I0317 17:02:37.356453 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:02:37 crc kubenswrapper[4767]: E0317 17:02:37.357294 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:02:37 crc kubenswrapper[4767]: I0317 17:02:37.943560 4767 scope.go:117] "RemoveContainer" containerID="7946dbc889642db464ce137528a6663b61012af67d1fc270514467ca79680c66" Mar 17 17:02:40 crc kubenswrapper[4767]: I0317 17:02:40.642721 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 17:02:40 crc kubenswrapper[4767]: I0317 17:02:40.704832 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v5jdf" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.519078 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qc7rs/must-gather-c55jv"] Mar 17 17:02:45 crc kubenswrapper[4767]: E0317 17:02:45.556538 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38f4981-a828-4985-9162-9597fd4dfc6d" containerName="keystone-cron" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.556619 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38f4981-a828-4985-9162-9597fd4dfc6d" containerName="keystone-cron" Mar 17 17:02:45 crc kubenswrapper[4767]: E0317 17:02:45.556767 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00fe51b-6938-436c-ae5b-f961f538af6b" containerName="oc" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.556774 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00fe51b-6938-436c-ae5b-f961f538af6b" containerName="oc" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.559317 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d38f4981-a828-4985-9162-9597fd4dfc6d" containerName="keystone-cron" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.559375 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00fe51b-6938-436c-ae5b-f961f538af6b" containerName="oc" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.567597 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qc7rs/must-gather-c55jv"] Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.567728 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qc7rs/must-gather-c55jv" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.572998 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qc7rs"/"kube-root-ca.crt" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.573270 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qc7rs"/"default-dockercfg-967nr" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.578319 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qc7rs"/"openshift-service-ca.crt" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.693465 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9785\" (UniqueName: \"kubernetes.io/projected/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-kube-api-access-s9785\") pod \"must-gather-c55jv\" (UID: \"90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd\") " pod="openshift-must-gather-qc7rs/must-gather-c55jv" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.693596 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-must-gather-output\") pod \"must-gather-c55jv\" (UID: \"90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd\") " pod="openshift-must-gather-qc7rs/must-gather-c55jv" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.796451 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9785\" (UniqueName: \"kubernetes.io/projected/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-kube-api-access-s9785\") pod \"must-gather-c55jv\" (UID: \"90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd\") " pod="openshift-must-gather-qc7rs/must-gather-c55jv" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.796542 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-must-gather-output\") pod \"must-gather-c55jv\" (UID: \"90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd\") " pod="openshift-must-gather-qc7rs/must-gather-c55jv" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.797725 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-must-gather-output\") pod \"must-gather-c55jv\" (UID: \"90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd\") " pod="openshift-must-gather-qc7rs/must-gather-c55jv" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.836116 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9785\" (UniqueName: \"kubernetes.io/projected/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-kube-api-access-s9785\") pod \"must-gather-c55jv\" (UID: \"90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd\") " pod="openshift-must-gather-qc7rs/must-gather-c55jv" Mar 17 17:02:45 crc kubenswrapper[4767]: I0317 17:02:45.921509 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qc7rs/must-gather-c55jv" Mar 17 17:02:47 crc kubenswrapper[4767]: I0317 17:02:47.045036 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qc7rs/must-gather-c55jv"] Mar 17 17:02:47 crc kubenswrapper[4767]: W0317 17:02:47.069884 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90cd47b1_ea89_4cfc_8cb3_d5b967d6cfbd.slice/crio-45a1021cada625ecf7690bb3ada011f6c7cc5bd5d74facf0bfd14bdd729f00bf WatchSource:0}: Error finding container 45a1021cada625ecf7690bb3ada011f6c7cc5bd5d74facf0bfd14bdd729f00bf: Status 404 returned error can't find the container with id 45a1021cada625ecf7690bb3ada011f6c7cc5bd5d74facf0bfd14bdd729f00bf Mar 17 17:02:47 crc kubenswrapper[4767]: I0317 17:02:47.626680 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qc7rs/must-gather-c55jv" event={"ID":"90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd","Type":"ContainerStarted","Data":"45a1021cada625ecf7690bb3ada011f6c7cc5bd5d74facf0bfd14bdd729f00bf"} Mar 17 17:02:51 crc kubenswrapper[4767]: I0317 17:02:51.356484 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:02:51 crc kubenswrapper[4767]: E0317 17:02:51.357720 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:02:59 crc kubenswrapper[4767]: I0317 17:02:59.775670 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mr7jq"] Mar 17 17:02:59 crc kubenswrapper[4767]: I0317 17:02:59.781737 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:02:59 crc kubenswrapper[4767]: I0317 17:02:59.800869 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mr7jq"] Mar 17 17:02:59 crc kubenswrapper[4767]: I0317 17:02:59.893152 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-utilities\") pod \"redhat-operators-mr7jq\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:02:59 crc kubenswrapper[4767]: I0317 17:02:59.893231 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-catalog-content\") pod \"redhat-operators-mr7jq\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:02:59 crc kubenswrapper[4767]: I0317 17:02:59.893403 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw94q\" (UniqueName: \"kubernetes.io/projected/c473466b-2d30-40bd-9e0a-54870b46519f-kube-api-access-fw94q\") pod \"redhat-operators-mr7jq\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:02:59 crc kubenswrapper[4767]: I0317 17:02:59.995795 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-utilities\") pod \"redhat-operators-mr7jq\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:02:59 crc kubenswrapper[4767]: I0317 17:02:59.995873 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-catalog-content\") pod \"redhat-operators-mr7jq\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:02:59 crc kubenswrapper[4767]: I0317 17:02:59.996193 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw94q\" (UniqueName: \"kubernetes.io/projected/c473466b-2d30-40bd-9e0a-54870b46519f-kube-api-access-fw94q\") pod \"redhat-operators-mr7jq\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:02:59 crc kubenswrapper[4767]: I0317 17:02:59.997291 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-catalog-content\") pod \"redhat-operators-mr7jq\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:02:59 crc kubenswrapper[4767]: I0317 17:02:59.997419 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-utilities\") pod \"redhat-operators-mr7jq\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:03:02 crc kubenswrapper[4767]: I0317 17:03:02.355404 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:03:02 crc kubenswrapper[4767]: E0317 17:03:02.356311 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:03:09 crc kubenswrapper[4767]: I0317 17:03:09.660437 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw94q\" (UniqueName: \"kubernetes.io/projected/c473466b-2d30-40bd-9e0a-54870b46519f-kube-api-access-fw94q\") pod \"redhat-operators-mr7jq\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:03:09 crc kubenswrapper[4767]: I0317 17:03:09.706032 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:03:09 crc kubenswrapper[4767]: I0317 17:03:09.712786 4767 trace.go:236] Trace[1663554849]: "Calculate volume metrics of storage for pod openshift-logging/logging-loki-compactor-0" (17-Mar-2026 17:03:03.164) (total time: 6545ms): Mar 17 17:03:09 crc kubenswrapper[4767]: Trace[1663554849]: [6.545544333s] [6.545544333s] END Mar 17 17:03:12 crc kubenswrapper[4767]: E0317 17:03:12.762016 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-must-gather:latest" Mar 17 17:03:12 crc kubenswrapper[4767]: E0317 17:03:12.767233 4767 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 17 17:03:12 crc kubenswrapper[4767]: container &Container{Name:gather,Image:quay.io/openstack-k8s-operators/openstack-must-gather:latest,Command:[/bin/bash -c if command -v setsid >/dev/null 2>&1 && command -v ps >/dev/null 2>&1 && command -v pkill >/dev/null 2>&1; then Mar 17 17:03:12 crc kubenswrapper[4767]: HAVE_SESSION_TOOLS=true Mar 17 17:03:12 crc kubenswrapper[4767]: else Mar 17 17:03:12 crc kubenswrapper[4767]: HAVE_SESSION_TOOLS=false Mar 17 17:03:12 crc kubenswrapper[4767]: fi Mar 17 17:03:12 crc kubenswrapper[4767]: Mar 17 17:03:12 crc kubenswrapper[4767]: Mar 17 17:03:12 crc kubenswrapper[4767]: echo "[disk usage checker] Started" Mar 17 17:03:12 crc kubenswrapper[4767]: target_dir="/must-gather" Mar 17 17:03:12 crc kubenswrapper[4767]: usage_percentage_limit="80" Mar 17 17:03:12 crc kubenswrapper[4767]: while true; do Mar 17 17:03:12 crc kubenswrapper[4767]: usage_percentage=$(df -P "$target_dir" | awk 'NR==2 {print $5}' | sed 's/%//') Mar 17 17:03:12 crc kubenswrapper[4767]: echo "[disk usage checker] Volume usage percentage: current = ${usage_percentage} ; allowed = ${usage_percentage_limit}" Mar 17 17:03:12 crc kubenswrapper[4767]: if [ "$usage_percentage" -gt "$usage_percentage_limit" ]; then Mar 17 17:03:12 crc kubenswrapper[4767]: echo "[disk usage checker] Disk usage exceeds the volume percentage of ${usage_percentage_limit} for mounted directory, terminating..." Mar 17 17:03:12 crc kubenswrapper[4767]: if [ "$HAVE_SESSION_TOOLS" = "true" ]; then Mar 17 17:03:12 crc kubenswrapper[4767]: ps -o sess --no-headers | sort -u | while read sid; do Mar 17 17:03:12 crc kubenswrapper[4767]: [[ "$sid" -eq "${$}" ]] && continue Mar 17 17:03:12 crc kubenswrapper[4767]: pkill --signal SIGKILL --session "$sid" Mar 17 17:03:12 crc kubenswrapper[4767]: done Mar 17 17:03:12 crc kubenswrapper[4767]: else Mar 17 17:03:12 crc kubenswrapper[4767]: kill 0 Mar 17 17:03:12 crc kubenswrapper[4767]: fi Mar 17 17:03:12 crc kubenswrapper[4767]: exit 1 Mar 17 17:03:12 crc kubenswrapper[4767]: fi Mar 17 17:03:12 crc kubenswrapper[4767]: sleep 5 Mar 17 17:03:12 crc kubenswrapper[4767]: done & if [ "$HAVE_SESSION_TOOLS" = "true" ]; then Mar 17 17:03:12 crc kubenswrapper[4767]: setsid -w bash <<-MUSTGATHER_EOF Mar 17 17:03:12 crc kubenswrapper[4767]: ADDITIONAL_NAMESPACES=kuttl,openshift-storage,openshift-marketplace,openshift-operators,sushy-emulator,tobiko OPENSTACK_DATABASES=ALL SOS_EDPM=all OMC=False SOS_DECOMPRESS=0 gather Mar 17 17:03:12 crc kubenswrapper[4767]: MUSTGATHER_EOF Mar 17 17:03:12 crc kubenswrapper[4767]: else Mar 17 17:03:12 crc kubenswrapper[4767]: ADDITIONAL_NAMESPACES=kuttl,openshift-storage,openshift-marketplace,openshift-operators,sushy-emulator,tobiko OPENSTACK_DATABASES=ALL SOS_EDPM=all OMC=False SOS_DECOMPRESS=0 gather Mar 17 17:03:12 crc kubenswrapper[4767]: fi; sync && echo 'Caches written to disk'],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:must-gather-output,ReadOnly:false,MountPath:/must-gather,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s9785,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod must-gather-c55jv_openshift-must-gather-qc7rs(90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Mar 17 17:03:12 crc kubenswrapper[4767]: > logger="UnhandledError" Mar 17 17:03:12 crc kubenswrapper[4767]: E0317 17:03:12.773788 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"gather\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"copy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-must-gather:latest\\\"\"]" pod="openshift-must-gather-qc7rs/must-gather-c55jv" podUID="90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd" Mar 17 17:03:13 crc kubenswrapper[4767]: E0317 17:03:13.241919 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"gather\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-must-gather:latest\\\"\", failed to \"StartContainer\" for \"copy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-must-gather:latest\\\"\"]" pod="openshift-must-gather-qc7rs/must-gather-c55jv" podUID="90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd" Mar 17 17:03:13 crc kubenswrapper[4767]: I0317 17:03:13.463815 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mr7jq"] Mar 17 17:03:14 crc kubenswrapper[4767]: I0317 17:03:14.249995 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mr7jq" event={"ID":"c473466b-2d30-40bd-9e0a-54870b46519f","Type":"ContainerStarted","Data":"5833dbf63aa79c0ff82499ebed540373ced1763242aa97010c15f5b8859f3b8b"} Mar 17 17:03:15 crc kubenswrapper[4767]: I0317 17:03:15.263841 4767 generic.go:334] "Generic (PLEG): container finished" podID="c473466b-2d30-40bd-9e0a-54870b46519f" containerID="f816988dca421de60043998b9631c693b9f143c7e255efe9d87089ba5ae2bb6c" exitCode=0 Mar 17 17:03:15 crc kubenswrapper[4767]: I0317 17:03:15.263972 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mr7jq" event={"ID":"c473466b-2d30-40bd-9e0a-54870b46519f","Type":"ContainerDied","Data":"f816988dca421de60043998b9631c693b9f143c7e255efe9d87089ba5ae2bb6c"} Mar 17 17:03:15 crc kubenswrapper[4767]: I0317 17:03:15.268144 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 17:03:16 crc kubenswrapper[4767]: I0317 17:03:16.354414 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:03:16 crc kubenswrapper[4767]: E0317 17:03:16.355084 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:03:17 crc kubenswrapper[4767]: I0317 17:03:17.293727 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mr7jq" event={"ID":"c473466b-2d30-40bd-9e0a-54870b46519f","Type":"ContainerStarted","Data":"07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b"} Mar 17 17:03:20 crc kubenswrapper[4767]: I0317 17:03:20.566901 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qc7rs/must-gather-c55jv"] Mar 17 17:03:20 crc kubenswrapper[4767]: I0317 17:03:20.580697 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qc7rs/must-gather-c55jv"] Mar 17 17:03:21 crc kubenswrapper[4767]: I0317 17:03:21.059716 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qc7rs/must-gather-c55jv" Mar 17 17:03:21 crc kubenswrapper[4767]: I0317 17:03:21.180338 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-must-gather-output\") pod \"90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd\" (UID: \"90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd\") " Mar 17 17:03:21 crc kubenswrapper[4767]: I0317 17:03:21.180714 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9785\" (UniqueName: \"kubernetes.io/projected/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-kube-api-access-s9785\") pod \"90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd\" (UID: \"90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd\") " Mar 17 17:03:21 crc kubenswrapper[4767]: I0317 17:03:21.182289 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd" (UID: "90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 17:03:21 crc kubenswrapper[4767]: I0317 17:03:21.192531 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-kube-api-access-s9785" (OuterVolumeSpecName: "kube-api-access-s9785") pod "90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd" (UID: "90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd"). InnerVolumeSpecName "kube-api-access-s9785". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 17:03:21 crc kubenswrapper[4767]: I0317 17:03:21.284381 4767 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 17 17:03:21 crc kubenswrapper[4767]: I0317 17:03:21.284424 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9785\" (UniqueName: \"kubernetes.io/projected/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd-kube-api-access-s9785\") on node \"crc\" DevicePath \"\"" Mar 17 17:03:21 crc kubenswrapper[4767]: I0317 17:03:21.373263 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd" path="/var/lib/kubelet/pods/90cd47b1-ea89-4cfc-8cb3-d5b967d6cfbd/volumes" Mar 17 17:03:21 crc kubenswrapper[4767]: I0317 17:03:21.457943 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qc7rs/must-gather-c55jv" Mar 17 17:03:22 crc kubenswrapper[4767]: I0317 17:03:22.473207 4767 generic.go:334] "Generic (PLEG): container finished" podID="c473466b-2d30-40bd-9e0a-54870b46519f" containerID="07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b" exitCode=0 Mar 17 17:03:22 crc kubenswrapper[4767]: I0317 17:03:22.473462 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mr7jq" event={"ID":"c473466b-2d30-40bd-9e0a-54870b46519f","Type":"ContainerDied","Data":"07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b"} Mar 17 17:03:23 crc kubenswrapper[4767]: I0317 17:03:23.490473 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mr7jq" event={"ID":"c473466b-2d30-40bd-9e0a-54870b46519f","Type":"ContainerStarted","Data":"6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07"} Mar 17 17:03:23 crc kubenswrapper[4767]: I0317 17:03:23.521365 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mr7jq" podStartSLOduration=16.855556378 podStartE2EDuration="24.521332667s" podCreationTimestamp="2026-03-17 17:02:59 +0000 UTC" firstStartedPulling="2026-03-17 17:03:15.266813427 +0000 UTC m=+5186.680129474" lastFinishedPulling="2026-03-17 17:03:22.932589716 +0000 UTC m=+5194.345905763" observedRunningTime="2026-03-17 17:03:23.519249668 +0000 UTC m=+5194.932565745" watchObservedRunningTime="2026-03-17 17:03:23.521332667 +0000 UTC m=+5194.934648714" Mar 17 17:03:28 crc kubenswrapper[4767]: I0317 17:03:28.355807 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:03:28 crc kubenswrapper[4767]: E0317 17:03:28.356821 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:03:29 crc kubenswrapper[4767]: I0317 17:03:29.707736 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:03:29 crc kubenswrapper[4767]: I0317 17:03:29.708129 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:03:32 crc kubenswrapper[4767]: I0317 17:03:32.016923 4767 trace.go:236] Trace[259508762]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-1" (17-Mar-2026 17:03:29.334) (total time: 2682ms): Mar 17 17:03:32 crc kubenswrapper[4767]: Trace[259508762]: [2.682481927s] [2.682481927s] END Mar 17 17:03:33 crc kubenswrapper[4767]: I0317 17:03:33.103338 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mr7jq" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" containerName="registry-server" probeResult="failure" output=< Mar 17 17:03:33 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:03:33 crc kubenswrapper[4767]: > Mar 17 17:03:40 crc kubenswrapper[4767]: I0317 17:03:40.354886 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:03:40 crc kubenswrapper[4767]: E0317 17:03:40.355733 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:03:40 crc kubenswrapper[4767]: I0317 17:03:40.759299 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mr7jq" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" containerName="registry-server" probeResult="failure" output=< Mar 17 17:03:40 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:03:40 crc kubenswrapper[4767]: > Mar 17 17:03:50 crc kubenswrapper[4767]: I0317 17:03:50.787974 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mr7jq" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" containerName="registry-server" probeResult="failure" output=< Mar 17 17:03:50 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Mar 17 17:03:50 crc kubenswrapper[4767]: > Mar 17 17:03:51 crc kubenswrapper[4767]: I0317 17:03:51.355772 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:03:51 crc kubenswrapper[4767]: E0317 17:03:51.356331 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:03:59 crc kubenswrapper[4767]: I0317 17:03:59.764081 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:03:59 crc kubenswrapper[4767]: I0317 17:03:59.822688 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:04:00 crc kubenswrapper[4767]: I0317 17:04:00.022136 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mr7jq"] Mar 17 17:04:00 crc kubenswrapper[4767]: I0317 17:04:00.179406 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562784-ddkrk"] Mar 17 17:04:00 crc kubenswrapper[4767]: I0317 17:04:00.181665 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562784-ddkrk" Mar 17 17:04:00 crc kubenswrapper[4767]: I0317 17:04:00.186566 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 17:04:00 crc kubenswrapper[4767]: I0317 17:04:00.186989 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 17:04:00 crc kubenswrapper[4767]: I0317 17:04:00.187297 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 17:04:00 crc kubenswrapper[4767]: I0317 17:04:00.198378 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562784-ddkrk"] Mar 17 17:04:00 crc kubenswrapper[4767]: I0317 17:04:00.287160 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2ggv\" (UniqueName: \"kubernetes.io/projected/0ce412e1-5e28-4a16-88e9-72c907f59d53-kube-api-access-k2ggv\") pod \"auto-csr-approver-29562784-ddkrk\" (UID: \"0ce412e1-5e28-4a16-88e9-72c907f59d53\") " pod="openshift-infra/auto-csr-approver-29562784-ddkrk" Mar 17 17:04:00 crc kubenswrapper[4767]: I0317 17:04:00.390288 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2ggv\" (UniqueName: \"kubernetes.io/projected/0ce412e1-5e28-4a16-88e9-72c907f59d53-kube-api-access-k2ggv\") pod \"auto-csr-approver-29562784-ddkrk\" (UID: \"0ce412e1-5e28-4a16-88e9-72c907f59d53\") " pod="openshift-infra/auto-csr-approver-29562784-ddkrk" Mar 17 17:04:01 crc kubenswrapper[4767]: I0317 17:04:01.049261 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mr7jq" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" containerName="registry-server" containerID="cri-o://6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07" gracePeriod=2 Mar 17 17:04:01 crc kubenswrapper[4767]: I0317 17:04:01.150248 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2ggv\" (UniqueName: \"kubernetes.io/projected/0ce412e1-5e28-4a16-88e9-72c907f59d53-kube-api-access-k2ggv\") pod \"auto-csr-approver-29562784-ddkrk\" (UID: \"0ce412e1-5e28-4a16-88e9-72c907f59d53\") " pod="openshift-infra/auto-csr-approver-29562784-ddkrk" Mar 17 17:04:01 crc kubenswrapper[4767]: I0317 17:04:01.415597 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562784-ddkrk" Mar 17 17:04:01 crc kubenswrapper[4767]: I0317 17:04:01.989899 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.065811 4767 generic.go:334] "Generic (PLEG): container finished" podID="c473466b-2d30-40bd-9e0a-54870b46519f" containerID="6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07" exitCode=0 Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.065875 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mr7jq" event={"ID":"c473466b-2d30-40bd-9e0a-54870b46519f","Type":"ContainerDied","Data":"6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07"} Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.065912 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mr7jq" event={"ID":"c473466b-2d30-40bd-9e0a-54870b46519f","Type":"ContainerDied","Data":"5833dbf63aa79c0ff82499ebed540373ced1763242aa97010c15f5b8859f3b8b"} Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.065932 4767 scope.go:117] "RemoveContainer" containerID="6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.066145 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mr7jq" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.114362 4767 scope.go:117] "RemoveContainer" containerID="07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.123940 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562784-ddkrk"] Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.137979 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw94q\" (UniqueName: \"kubernetes.io/projected/c473466b-2d30-40bd-9e0a-54870b46519f-kube-api-access-fw94q\") pod \"c473466b-2d30-40bd-9e0a-54870b46519f\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.138059 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-utilities\") pod \"c473466b-2d30-40bd-9e0a-54870b46519f\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.138160 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-catalog-content\") pod \"c473466b-2d30-40bd-9e0a-54870b46519f\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.138981 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-utilities" (OuterVolumeSpecName: "utilities") pod "c473466b-2d30-40bd-9e0a-54870b46519f" (UID: "c473466b-2d30-40bd-9e0a-54870b46519f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.139615 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.142783 4767 scope.go:117] "RemoveContainer" containerID="f816988dca421de60043998b9631c693b9f143c7e255efe9d87089ba5ae2bb6c" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.145101 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c473466b-2d30-40bd-9e0a-54870b46519f-kube-api-access-fw94q" (OuterVolumeSpecName: "kube-api-access-fw94q") pod "c473466b-2d30-40bd-9e0a-54870b46519f" (UID: "c473466b-2d30-40bd-9e0a-54870b46519f"). InnerVolumeSpecName "kube-api-access-fw94q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.168558 4767 scope.go:117] "RemoveContainer" containerID="6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07" Mar 17 17:04:02 crc kubenswrapper[4767]: E0317 17:04:02.169127 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07\": container with ID starting with 6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07 not found: ID does not exist" containerID="6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.169194 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07"} err="failed to get container status \"6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07\": rpc error: code = NotFound desc = could not find container \"6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07\": container with ID starting with 6eae0a1983fda75bf751e7616e07338bea7f9eab2db2b8ecb690dc57849a1e07 not found: ID does not exist" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.169233 4767 scope.go:117] "RemoveContainer" containerID="07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b" Mar 17 17:04:02 crc kubenswrapper[4767]: E0317 17:04:02.169884 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b\": container with ID starting with 07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b not found: ID does not exist" containerID="07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.170020 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b"} err="failed to get container status \"07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b\": rpc error: code = NotFound desc = could not find container \"07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b\": container with ID starting with 07ef710db0f8c90981bd01c250b31b7cae5001d9999c60d0caf06135d04b038b not found: ID does not exist" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.170054 4767 scope.go:117] "RemoveContainer" containerID="f816988dca421de60043998b9631c693b9f143c7e255efe9d87089ba5ae2bb6c" Mar 17 17:04:02 crc kubenswrapper[4767]: E0317 17:04:02.170390 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f816988dca421de60043998b9631c693b9f143c7e255efe9d87089ba5ae2bb6c\": container with ID starting with f816988dca421de60043998b9631c693b9f143c7e255efe9d87089ba5ae2bb6c not found: ID does not exist" containerID="f816988dca421de60043998b9631c693b9f143c7e255efe9d87089ba5ae2bb6c" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.170413 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f816988dca421de60043998b9631c693b9f143c7e255efe9d87089ba5ae2bb6c"} err="failed to get container status \"f816988dca421de60043998b9631c693b9f143c7e255efe9d87089ba5ae2bb6c\": rpc error: code = NotFound desc = could not find container \"f816988dca421de60043998b9631c693b9f143c7e255efe9d87089ba5ae2bb6c\": container with ID starting with f816988dca421de60043998b9631c693b9f143c7e255efe9d87089ba5ae2bb6c not found: ID does not exist" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.242770 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw94q\" (UniqueName: \"kubernetes.io/projected/c473466b-2d30-40bd-9e0a-54870b46519f-kube-api-access-fw94q\") on node \"crc\" DevicePath \"\"" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.344217 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c473466b-2d30-40bd-9e0a-54870b46519f" (UID: "c473466b-2d30-40bd-9e0a-54870b46519f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.344924 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-catalog-content\") pod \"c473466b-2d30-40bd-9e0a-54870b46519f\" (UID: \"c473466b-2d30-40bd-9e0a-54870b46519f\") " Mar 17 17:04:02 crc kubenswrapper[4767]: W0317 17:04:02.346561 4767 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/c473466b-2d30-40bd-9e0a-54870b46519f/volumes/kubernetes.io~empty-dir/catalog-content Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.346587 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c473466b-2d30-40bd-9e0a-54870b46519f" (UID: "c473466b-2d30-40bd-9e0a-54870b46519f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.414062 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mr7jq"] Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.427442 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mr7jq"] Mar 17 17:04:02 crc kubenswrapper[4767]: I0317 17:04:02.449801 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c473466b-2d30-40bd-9e0a-54870b46519f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 17:04:03 crc kubenswrapper[4767]: I0317 17:04:03.077235 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562784-ddkrk" event={"ID":"0ce412e1-5e28-4a16-88e9-72c907f59d53","Type":"ContainerStarted","Data":"17fb68c29dc89657085f7ba07595ee501039af6b3b242522bcc5d7f5c14b8340"} Mar 17 17:04:03 crc kubenswrapper[4767]: I0317 17:04:03.373696 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" path="/var/lib/kubelet/pods/c473466b-2d30-40bd-9e0a-54870b46519f/volumes" Mar 17 17:04:04 crc kubenswrapper[4767]: I0317 17:04:04.355964 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:04:04 crc kubenswrapper[4767]: E0317 17:04:04.356726 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:04:05 crc kubenswrapper[4767]: I0317 17:04:05.108122 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562784-ddkrk" event={"ID":"0ce412e1-5e28-4a16-88e9-72c907f59d53","Type":"ContainerStarted","Data":"3273f7f35c71ab78dcc36fe24a5b1635c1b4f1e6169b8124c0676998fa114518"} Mar 17 17:04:05 crc kubenswrapper[4767]: I0317 17:04:05.131373 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562784-ddkrk" podStartSLOduration=4.117970537 podStartE2EDuration="5.131348039s" podCreationTimestamp="2026-03-17 17:04:00 +0000 UTC" firstStartedPulling="2026-03-17 17:04:02.143510326 +0000 UTC m=+5233.556826373" lastFinishedPulling="2026-03-17 17:04:03.156887828 +0000 UTC m=+5234.570203875" observedRunningTime="2026-03-17 17:04:05.123115685 +0000 UTC m=+5236.536431732" watchObservedRunningTime="2026-03-17 17:04:05.131348039 +0000 UTC m=+5236.544664086" Mar 17 17:04:06 crc kubenswrapper[4767]: I0317 17:04:06.123196 4767 generic.go:334] "Generic (PLEG): container finished" podID="0ce412e1-5e28-4a16-88e9-72c907f59d53" containerID="3273f7f35c71ab78dcc36fe24a5b1635c1b4f1e6169b8124c0676998fa114518" exitCode=0 Mar 17 17:04:06 crc kubenswrapper[4767]: I0317 17:04:06.123300 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562784-ddkrk" event={"ID":"0ce412e1-5e28-4a16-88e9-72c907f59d53","Type":"ContainerDied","Data":"3273f7f35c71ab78dcc36fe24a5b1635c1b4f1e6169b8124c0676998fa114518"} Mar 17 17:04:07 crc kubenswrapper[4767]: I0317 17:04:07.617273 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562784-ddkrk" Mar 17 17:04:07 crc kubenswrapper[4767]: I0317 17:04:07.722372 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2ggv\" (UniqueName: \"kubernetes.io/projected/0ce412e1-5e28-4a16-88e9-72c907f59d53-kube-api-access-k2ggv\") pod \"0ce412e1-5e28-4a16-88e9-72c907f59d53\" (UID: \"0ce412e1-5e28-4a16-88e9-72c907f59d53\") " Mar 17 17:04:07 crc kubenswrapper[4767]: I0317 17:04:07.730751 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ce412e1-5e28-4a16-88e9-72c907f59d53-kube-api-access-k2ggv" (OuterVolumeSpecName: "kube-api-access-k2ggv") pod "0ce412e1-5e28-4a16-88e9-72c907f59d53" (UID: "0ce412e1-5e28-4a16-88e9-72c907f59d53"). InnerVolumeSpecName "kube-api-access-k2ggv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 17:04:07 crc kubenswrapper[4767]: I0317 17:04:07.826573 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2ggv\" (UniqueName: \"kubernetes.io/projected/0ce412e1-5e28-4a16-88e9-72c907f59d53-kube-api-access-k2ggv\") on node \"crc\" DevicePath \"\"" Mar 17 17:04:08 crc kubenswrapper[4767]: I0317 17:04:08.149557 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562784-ddkrk" event={"ID":"0ce412e1-5e28-4a16-88e9-72c907f59d53","Type":"ContainerDied","Data":"17fb68c29dc89657085f7ba07595ee501039af6b3b242522bcc5d7f5c14b8340"} Mar 17 17:04:08 crc kubenswrapper[4767]: I0317 17:04:08.149611 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17fb68c29dc89657085f7ba07595ee501039af6b3b242522bcc5d7f5c14b8340" Mar 17 17:04:08 crc kubenswrapper[4767]: I0317 17:04:08.149643 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562784-ddkrk" Mar 17 17:04:08 crc kubenswrapper[4767]: I0317 17:04:08.212518 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562778-5cxf5"] Mar 17 17:04:08 crc kubenswrapper[4767]: I0317 17:04:08.225162 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562778-5cxf5"] Mar 17 17:04:09 crc kubenswrapper[4767]: I0317 17:04:09.368820 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f71df388-7598-46d0-82f6-5a6aa85d90ad" path="/var/lib/kubelet/pods/f71df388-7598-46d0-82f6-5a6aa85d90ad/volumes" Mar 17 17:04:16 crc kubenswrapper[4767]: I0317 17:04:16.355468 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:04:16 crc kubenswrapper[4767]: E0317 17:04:16.356305 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:04:29 crc kubenswrapper[4767]: I0317 17:04:29.363504 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:04:29 crc kubenswrapper[4767]: E0317 17:04:29.365560 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:04:38 crc kubenswrapper[4767]: I0317 17:04:38.212318 4767 scope.go:117] "RemoveContainer" containerID="f752ad2aed374732f44f8dca7dc468c6deeb94183f06dfb477226dabf512e86a" Mar 17 17:04:41 crc kubenswrapper[4767]: I0317 17:04:41.355442 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:04:41 crc kubenswrapper[4767]: E0317 17:04:41.356369 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:04:52 crc kubenswrapper[4767]: I0317 17:04:52.354462 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:04:52 crc kubenswrapper[4767]: E0317 17:04:52.355466 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:05:07 crc kubenswrapper[4767]: I0317 17:05:07.355288 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:05:07 crc kubenswrapper[4767]: E0317 17:05:07.356408 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:05:22 crc kubenswrapper[4767]: I0317 17:05:22.355533 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:05:22 crc kubenswrapper[4767]: E0317 17:05:22.357768 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:05:34 crc kubenswrapper[4767]: I0317 17:05:34.356690 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:05:34 crc kubenswrapper[4767]: E0317 17:05:34.358871 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:05:47 crc kubenswrapper[4767]: I0317 17:05:47.358306 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:05:47 crc kubenswrapper[4767]: E0317 17:05:47.359692 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:05:58 crc kubenswrapper[4767]: I0317 17:05:58.355770 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:05:58 crc kubenswrapper[4767]: E0317 17:05:58.356615 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.154826 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562786-vccr4"] Mar 17 17:06:00 crc kubenswrapper[4767]: E0317 17:06:00.155801 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" containerName="extract-utilities" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.155820 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" containerName="extract-utilities" Mar 17 17:06:00 crc kubenswrapper[4767]: E0317 17:06:00.155853 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" containerName="extract-content" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.155859 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" containerName="extract-content" Mar 17 17:06:00 crc kubenswrapper[4767]: E0317 17:06:00.155878 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" containerName="registry-server" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.155884 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" containerName="registry-server" Mar 17 17:06:00 crc kubenswrapper[4767]: E0317 17:06:00.155908 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce412e1-5e28-4a16-88e9-72c907f59d53" containerName="oc" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.155914 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce412e1-5e28-4a16-88e9-72c907f59d53" containerName="oc" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.156218 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ce412e1-5e28-4a16-88e9-72c907f59d53" containerName="oc" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.156259 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c473466b-2d30-40bd-9e0a-54870b46519f" containerName="registry-server" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.158359 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562786-vccr4" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.161910 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-w2l2m" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.162443 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.169500 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.170922 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562786-vccr4"] Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.256304 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p9kp\" (UniqueName: \"kubernetes.io/projected/b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45-kube-api-access-6p9kp\") pod \"auto-csr-approver-29562786-vccr4\" (UID: \"b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45\") " pod="openshift-infra/auto-csr-approver-29562786-vccr4" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.358731 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p9kp\" (UniqueName: \"kubernetes.io/projected/b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45-kube-api-access-6p9kp\") pod \"auto-csr-approver-29562786-vccr4\" (UID: \"b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45\") " pod="openshift-infra/auto-csr-approver-29562786-vccr4" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.385453 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p9kp\" (UniqueName: \"kubernetes.io/projected/b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45-kube-api-access-6p9kp\") pod \"auto-csr-approver-29562786-vccr4\" (UID: \"b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45\") " pod="openshift-infra/auto-csr-approver-29562786-vccr4" Mar 17 17:06:00 crc kubenswrapper[4767]: I0317 17:06:00.487382 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562786-vccr4" Mar 17 17:06:01 crc kubenswrapper[4767]: I0317 17:06:01.073014 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562786-vccr4"] Mar 17 17:06:01 crc kubenswrapper[4767]: I0317 17:06:01.679987 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562786-vccr4" event={"ID":"b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45","Type":"ContainerStarted","Data":"9285e64b0a019ebeb9d6469fa51367e151aab4f5721103f3d840b14665ab18b9"} Mar 17 17:06:03 crc kubenswrapper[4767]: I0317 17:06:03.704832 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562786-vccr4" event={"ID":"b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45","Type":"ContainerStarted","Data":"81ee15261b3f951c4a555065536fb5acb16a5072d3f8bc2d489a6bbbb258043e"} Mar 17 17:06:03 crc kubenswrapper[4767]: I0317 17:06:03.728000 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562786-vccr4" podStartSLOduration=2.370571624 podStartE2EDuration="3.727975682s" podCreationTimestamp="2026-03-17 17:06:00 +0000 UTC" firstStartedPulling="2026-03-17 17:06:01.076064789 +0000 UTC m=+5352.489380836" lastFinishedPulling="2026-03-17 17:06:02.433468847 +0000 UTC m=+5353.846784894" observedRunningTime="2026-03-17 17:06:03.720252803 +0000 UTC m=+5355.133568870" watchObservedRunningTime="2026-03-17 17:06:03.727975682 +0000 UTC m=+5355.141291729" Mar 17 17:06:04 crc kubenswrapper[4767]: I0317 17:06:04.730451 4767 generic.go:334] "Generic (PLEG): container finished" podID="b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45" containerID="81ee15261b3f951c4a555065536fb5acb16a5072d3f8bc2d489a6bbbb258043e" exitCode=0 Mar 17 17:06:04 crc kubenswrapper[4767]: I0317 17:06:04.730555 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562786-vccr4" event={"ID":"b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45","Type":"ContainerDied","Data":"81ee15261b3f951c4a555065536fb5acb16a5072d3f8bc2d489a6bbbb258043e"} Mar 17 17:06:13 crc kubenswrapper[4767]: I0317 17:06:13.358922 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:06:13 crc kubenswrapper[4767]: E0317 17:06:13.360053 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:06:24 crc kubenswrapper[4767]: I0317 17:06:24.354425 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:06:24 crc kubenswrapper[4767]: E0317 17:06:24.355491 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:06:26 crc kubenswrapper[4767]: I0317 17:06:26.854810 4767 trace.go:236] Trace[607482603]: "Calculate volume metrics of ovndbcluster-nb-etc-ovn for pod openstack/ovsdbserver-nb-0" (17-Mar-2026 17:06:11.625) (total time: 15229ms): Mar 17 17:06:26 crc kubenswrapper[4767]: Trace[607482603]: [15.229760396s] [15.229760396s] END Mar 17 17:06:26 crc kubenswrapper[4767]: I0317 17:06:26.855168 4767 trace.go:236] Trace[573892855]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-1" (17-Mar-2026 17:06:08.162) (total time: 18692ms): Mar 17 17:06:26 crc kubenswrapper[4767]: Trace[573892855]: [18.692283652s] [18.692283652s] END Mar 17 17:06:26 crc kubenswrapper[4767]: I0317 17:06:26.986110 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562786-vccr4" Mar 17 17:06:27 crc kubenswrapper[4767]: I0317 17:06:27.133722 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p9kp\" (UniqueName: \"kubernetes.io/projected/b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45-kube-api-access-6p9kp\") pod \"b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45\" (UID: \"b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45\") " Mar 17 17:06:27 crc kubenswrapper[4767]: I0317 17:06:27.166467 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45-kube-api-access-6p9kp" (OuterVolumeSpecName: "kube-api-access-6p9kp") pod "b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45" (UID: "b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45"). InnerVolumeSpecName "kube-api-access-6p9kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 17:06:27 crc kubenswrapper[4767]: I0317 17:06:27.229071 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562786-vccr4" event={"ID":"b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45","Type":"ContainerDied","Data":"9285e64b0a019ebeb9d6469fa51367e151aab4f5721103f3d840b14665ab18b9"} Mar 17 17:06:27 crc kubenswrapper[4767]: I0317 17:06:27.229120 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9285e64b0a019ebeb9d6469fa51367e151aab4f5721103f3d840b14665ab18b9" Mar 17 17:06:27 crc kubenswrapper[4767]: I0317 17:06:27.229207 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562786-vccr4" Mar 17 17:06:27 crc kubenswrapper[4767]: I0317 17:06:27.238832 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p9kp\" (UniqueName: \"kubernetes.io/projected/b0aa91c1-2cda-4a5e-bdcd-43e802ae0d45-kube-api-access-6p9kp\") on node \"crc\" DevicePath \"\"" Mar 17 17:06:28 crc kubenswrapper[4767]: I0317 17:06:28.096007 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562780-s2ljx"] Mar 17 17:06:28 crc kubenswrapper[4767]: I0317 17:06:28.112618 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562780-s2ljx"] Mar 17 17:06:29 crc kubenswrapper[4767]: I0317 17:06:29.371791 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af06aea-252b-458b-8180-2ec0bba5f0c2" path="/var/lib/kubelet/pods/1af06aea-252b-458b-8180-2ec0bba5f0c2/volumes" Mar 17 17:06:38 crc kubenswrapper[4767]: I0317 17:06:38.446971 4767 scope.go:117] "RemoveContainer" containerID="f0dab67d7e1a35ec5caea2c1b2615e5b4583111aac4e3ad25d3f29e990a838ee" Mar 17 17:06:39 crc kubenswrapper[4767]: I0317 17:06:39.370423 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:06:39 crc kubenswrapper[4767]: E0317 17:06:39.371122 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:06:50 crc kubenswrapper[4767]: I0317 17:06:50.356480 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:06:50 crc kubenswrapper[4767]: E0317 17:06:50.357437 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:07:01 crc kubenswrapper[4767]: I0317 17:07:01.358458 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:07:01 crc kubenswrapper[4767]: E0317 17:07:01.359520 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lhrtr_openshift-machine-config-operator(d680dc75-619e-41ae-90cb-2f58898a33ac)\"" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" podUID="d680dc75-619e-41ae-90cb-2f58898a33ac" Mar 17 17:07:15 crc kubenswrapper[4767]: I0317 17:07:15.355792 4767 scope.go:117] "RemoveContainer" containerID="1480b2fb1dd1f11de39acfa148bc0a45ee9f1675e5de38ef1a58dbbd28e58d6e" Mar 17 17:07:16 crc kubenswrapper[4767]: I0317 17:07:16.186977 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lhrtr" event={"ID":"d680dc75-619e-41ae-90cb-2f58898a33ac","Type":"ContainerStarted","Data":"b71c7351605b9775f5c3b413e2f3de93ed73ccde063f896865e91390108c8641"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515156305171024451 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015156305172017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015156272121016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015156272121015457 5ustar corecore